Improve model card: Add pipeline tag, OCR/Qwen tags, HF paper link, and sample usage
Browse filesThis PR enhances the model card for `olmOCR-7B-0725-FP8` by:
- Adding `pipeline_tag: image-to-text` to improve discoverability for OCR tasks.
- Adding relevant `tags: ocr, qwen` to further categorize the model based on its functionality and base architecture.
- Updating the "Paper" link in the Quick links section to point to the official Hugging Face Papers page ([https://huggingface.co/papers/2510.19817](https://huggingface.co/papers/2510.19817)) for better integration.
- Updating the introductory description to reflect the context of the `olmOCR 2` paper.
- Introducing a "Sample Usage" section with a bash code snippet directly from the GitHub README to help users quickly get started with local inference.
These changes will make the model card more informative and user-friendly on the Hugging Face Hub.
|
@@ -1,26 +1,29 @@
|
|
| 1 |
---
|
| 2 |
-
language:
|
| 3 |
-
- en
|
| 4 |
-
license: apache-2.0
|
| 5 |
-
datasets:
|
| 6 |
-
- allenai/olmOCR-mix-0225
|
| 7 |
base_model:
|
| 8 |
- Qwen/Qwen2.5-VL-7B-Instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
library_name: transformers
|
|
|
|
| 10 |
new_version: allenai/olmOCR-7B-0825-FP8
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
<img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
|
| 14 |
|
| 15 |
# olmOCR-7B-0725-FP8
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
-
This
|
| 20 |
-
[olmOCR-mix-0225](https://huggingface.co/datasets/allenai/olmOCR-mix-0225) dataset.
|
| 21 |
|
| 22 |
Quick links:
|
| 23 |
-
- ๐ [Paper](https://
|
| 24 |
- ๐ค [Dataset](https://huggingface.co/datasets/allenai/olmOCR-mix-0225)
|
| 25 |
- ๐ ๏ธ [Code](https://github.com/allenai/olmocr)
|
| 26 |
- ๐ฎ [Demo](https://olmocr.allenai.org/)
|
|
@@ -36,6 +39,19 @@ This model expects as input a single document image, rendered such that the long
|
|
| 36 |
The prompt must then contain the additional metadata from the document, and the easiest way to generate this
|
| 37 |
is to use the methods provided by the [olmOCR toolkit](https://github.com/allenai/olmocr).
|
| 38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
|
| 40 |
## License and use
|
| 41 |
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen2.5-VL-7B-Instruct
|
| 4 |
+
datasets:
|
| 5 |
+
- allenai/olmOCR-mix-0225
|
| 6 |
+
language:
|
| 7 |
+
- en
|
| 8 |
library_name: transformers
|
| 9 |
+
license: apache-2.0
|
| 10 |
new_version: allenai/olmOCR-7B-0825-FP8
|
| 11 |
+
pipeline_tag: image-to-text
|
| 12 |
+
tags:
|
| 13 |
+
- ocr
|
| 14 |
+
- qwen
|
| 15 |
---
|
| 16 |
|
| 17 |
<img alt="olmOCR Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmocr/olmocr.png" width="242px" style="margin-left:'auto' margin-right:'auto' display:'block'">
|
| 18 |
|
| 19 |
# olmOCR-7B-0725-FP8
|
| 20 |
|
| 21 |
+
This is a quantized (FP8) version of [olmOCR-7B-0725](https://huggingface.co/allenai/olmOCR-7B-0725), using llmcompressor. This model is part of the olmOCR family of powerful OCR systems for converting digitized print documents, like PDFs, into clean, naturally ordered plain text. The latest advancements in this family are presented in the paper [olmOCR 2: Unit Test Rewards for Document OCR](https://huggingface.co/papers/2510.19817).
|
| 22 |
|
| 23 |
+
This olmOCR model is fine-tuned from Qwen2.5-VL-7B-Instruct using the [olmOCR-mix-0225](https://huggingface.co/datasets/allenai/olmOCR-mix-0225) dataset.
|
|
|
|
| 24 |
|
| 25 |
Quick links:
|
| 26 |
+
- ๐ [Paper](https://huggingface.co/papers/2510.19817)
|
| 27 |
- ๐ค [Dataset](https://huggingface.co/datasets/allenai/olmOCR-mix-0225)
|
| 28 |
- ๐ ๏ธ [Code](https://github.com/allenai/olmocr)
|
| 29 |
- ๐ฎ [Demo](https://olmocr.allenai.org/)
|
|
|
|
| 39 |
The prompt must then contain the additional metadata from the document, and the easiest way to generate this
|
| 40 |
is to use the methods provided by the [olmOCR toolkit](https://github.com/allenai/olmocr).
|
| 41 |
|
| 42 |
+
## Sample Usage
|
| 43 |
+
|
| 44 |
+
For quick testing, try the [web demo](https://olmocr.allen.ai/). To run locally, a GPU is required. Here's an example of converting a single PDF using the `olmocr` pipeline, as provided in the GitHub repository:
|
| 45 |
+
|
| 46 |
+
```bash
|
| 47 |
+
# Download a sample PDF
|
| 48 |
+
curl -o olmocr-sample.pdf https://olmocr.allenai.org/papers/olmocr_3pg_sample.pdf
|
| 49 |
+
|
| 50 |
+
# Convert it to markdown
|
| 51 |
+
python -m olmocr.pipeline ./localworkspace --markdown --pdfs olmocr-sample.pdf
|
| 52 |
+
```
|
| 53 |
+
|
| 54 |
+
Results will be stored as markdown files inside of `./localworkspace/markdown/`.
|
| 55 |
|
| 56 |
## License and use
|
| 57 |
|