Install
openclaw skills install extracting-mistral-ocrExtracts text, tables, and images from PDFs (including scanned PDFs) using the Mistral OCR API. Use when user asks to OCR a PDF/image, extract text from a PDF, parse a scanned document, convert a PDF to Markdown, or extract structured fields from a document.
openclaw skills install extracting-mistral-ocrRun the bundled script to OCR a local PDF and write Markdown + JSON outputs:
python {baseDir}/scripts/mistral_ocr_extract.py --input path/to/file.pdf --out out/ocr
Output directory layout:
combined.md (all pages concatenated)pages/page-000.md (per-page markdown)raw_response.json (full OCR response)images/ (decoded embedded images, if requested)tables/ (separate tables, if requested)Pick input mode
file_id.document_url.Choose output fidelity (defaults are safe for RAG)
table_format=inline unless the user explicitly wants tables split out.--include-image-base64 when the user needs figures/diagrams extracted.--extract-header/--extract-footer if header/footer noise hurts downstream search.Run OCR
scripts/mistral_ocr_extract.py to produce a deterministic on-disk artefact set.(Optional) Structured extraction from the whole document
document_annotation in addition to page markdown.Example:
python {baseDir}/scripts/mistral_ocr_extract.py \
--input invoice.pdf \
--out out/invoice \
--annotation-prompt "Extract supplier_name, invoice_number, invoice_date (ISO-8601), currency, total_amount. Return JSON." \
--annotation-format json_object
document_url; upload instead.table_format=html for downstream parsing over brittle regex.MISTRAL_API_KEY: set it in the environment before running.--pages) or batch processing.references/mistral_ocr_api.mdreferences/output_mapping.mdreferences/annotation_prompts.md