I am looking for a small VLM/LLM which will run on mobile to help with extraction of data from Medical Lab Reports. I currently doing normal OCR and need LLM for verification of the data extracted and summarization.
u/Ecstatic-Tension-898
▲ 1 r/reactnative
u/Ecstatic-Tension-898 — 15 days ago
▲ 1 r/LocalLLM
I am working on a Mobile application which extracts data from Lab Reports. I already have traditional OCR extracting values but it makes mistakes, especially with different lab formats. I want to use a VLM to verify and correct what OCR gets wrong.
Here's what I have tested so far:
* **SmolVLM 500M:** runs great but it's too small. Returns "1.0" for literally every value, can describe the table but can't actually read it.
* **Qwen 2.5 VL 3B:** crashes from memory pressure
* **MedGemma 4B:** PaliGemma2 arch not supported by llama.cpp. Even text-only produces `<unused57>` garbage tokens.
Has anyone worked on similar use case?
u/Ecstatic-Tension-898 — 16 days ago