This is an automated archive made by the Lemmit Bot.
The original was posted on /r/machinelearning by /u/SouvikMandal on 2025-06-12 15:41:51+00:00.
We’re excited to share Nanonets-OCR-s, a powerful and lightweight (3B) VLM model that converts documents into clean, structured Markdown. This model is trained to understand document structure and content context (like tables, equations, images, plots, watermarks, checkboxes, etc.).
🔍 Key Features:
- LaTeX Equation Recognition Converts inline and block-level math into properly formatted LaTeX, distinguishing between
$...$
and$$...$$
. - Image Descriptions for LLMs Describes embedded images using structured
<img>
tags. Handles logos, charts, plots, and so on. - Signature Detection & Isolation Finds and tags signatures in scanned documents, outputting them in
<signature>
blocks. - Watermark Extraction Extracts watermark text and stores it within
<watermark>
tag for traceability. - Smart Checkbox & Radio Button Handling Converts checkboxes to Unicode symbols like ☑, ☒, and ☐ for reliable parsing in downstream apps.
- Complex Table Extraction Handles multi-row/column tables, preserving structure and outputting both Markdown and HTML formats.
Huggingface / GitHub / Try it out:
You must log in or register to comment.