diff --git a/docs/examples/vlm_pipeline_api_model.py b/docs/examples/vlm_pipeline_api_model.py index 01bbbb14..badcb769 100644 --- a/docs/examples/vlm_pipeline_api_model.py +++ b/docs/examples/vlm_pipeline_api_model.py @@ -198,7 +198,7 @@ def main(): # Example using the OlmOcr (dynamic prompt) model with Ollama: # (uncomment the following lines) # pipeline_options.vlm_options = ollama_olmocr_vlm_options( - # model="hf.co/allenai/olmOCR-7B-0225-preview-GGUF:Q8_0", + # model="hf.co/allenai/olmOCR-7B-0225-preview", # ) # Another possibility is using online services, e.g. watsonx.ai.