Merge branch 'vlm-dynamic-prompt' of https://github.com/shkarupa-alex/docling into vlm-dynamic-prompt

This commit is contained in:
Shkarupa Alex 2025-07-07 14:29:20 +03:00
commit 5d209f5db4

View File

@ -198,7 +198,7 @@ def main():
# Example using the OlmOcr (dynamic prompt) model with Ollama:
# (uncomment the following lines)
# pipeline_options.vlm_options = ollama_olmocr_vlm_options(
# model="hf.co/allenai/olmOCR-7B-0225-preview-GGUF:Q8_0",
# model="hf.co/allenai/olmOCR-7B-0225-preview",
# )
# Another possibility is using online services, e.g. watsonx.ai.