diff --git a/docs/examples/vlm_pipeline_api_model.py b/docs/examples/vlm_pipeline_api_model.py index badcb769..e198811f 100644 --- a/docs/examples/vlm_pipeline_api_model.py +++ b/docs/examples/vlm_pipeline_api_model.py @@ -198,7 +198,7 @@ def main(): # Example using the OlmOcr (dynamic prompt) model with Ollama: # (uncomment the following lines) # pipeline_options.vlm_options = ollama_olmocr_vlm_options( - # model="hf.co/allenai/olmOCR-7B-0225-preview", + # model="hf.co/lmstudio-community/olmOCR-7B-0225-preview-GGUF:Q8_0", # ) # Another possibility is using online services, e.g. watsonx.ai.