Intro

The model* is fine-tuned on Qwen2.5-3B-VL using a dataset of invoices and receipts. It can be used to extract text from the input and return the output in a specified JSon format.

*It is already merged with the LoRA layer and the original model. Be mindful of the input size to avoid a CUDA out-of-memory error.

Here is an example notebook of inference

For the LoRA params only, go to this repo

Usage:

from transformers import AutoModelForVision2Seq, AutoProcessor, AutoTokenizer

model = AutoModelForVision2Seq.from_pretrained(
    "callmeeric5/Qwen3B-Invoice-Receipt",
    device_map="cuda", #auto
    torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("callmeeric5/Qwen-3B-Invoice-Receipt")
processor = AutoProcessor.from_pretrained("callmeeric5/Qwen-3B-Invoice-Receipt")
Downloads last month
375
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train callmeeric5/Qwen3B-Invoice-Receipt

Space using callmeeric5/Qwen3B-Invoice-Receipt 1

Collection including callmeeric5/Qwen3B-Invoice-Receipt