This is a curiosity post/question.<p>I am trying to extract supplement nutrition data from photos of products. Over the last 24 hours, I've tried every popular OCR software and API, but none of them come even close OpenAI. With OpenAI Vision I was able to extract tabular data even from very obscure/low-res pictures. Meanwhile, OCR APIs offered by Microsoft, Amazon, Google, etc. and OpenSource softwares like tesseract, EasyOCR, etc. can barely extract the same information from screenshots.<p>What is so fundamentally different about how OpenAI extracts/interprets texts from the traditional OCR softwares?
Because it understands what it's looking at. No clearer way to put it.<p>If you compare scanning random garbage text, it will likely be as bad as the others, including humans.