Hacker News new | past | comments | ask | show | jobs | submit login

If you want to run locally you can look into this https://github.com/PaddlePaddle/PaddleOCR

https://andrejusb.blogspot.com/2024/03/optimizing-receipt-pr...

But I suggest that you just skip that and use gpt-4o. They aren't actually going to steal your data.

Sort through it to find anything with a credit card number or anything ahead time.

Or you could look into InternVL..

Or a combination of PaddleOCR first and then use a strong LLM via API, like gpt-4o or llama3 70b via together.ai

If you truly must do it locally, then if you have two 3090s or 4090s it might work out. Otherwise it the LLMs may not be smart enough to give good results.

Leaving out the details of your hardware makes it impossible to give good advice about running locally. Other than, it's not really necessary.




> But I suggest that you just skip that and use gpt-4o. They aren't actually going to steal your data.

Why do you have this confidence? Is it based on reading their TOS, and assuming they'll follow it?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: