vikp 3 days ago

I ran a partial benchmark against marker - https://github.com/VikParuchuri/marker .

Across 375 samples with LLM as a judge, mistral scores 4.32, and marker 4.41 . Marker can inference between 20 and 120 pages per second on an H100.

You can see the samples here - https://huggingface.co/datasets/datalab-to/marker_comparison... .

The code for the benchmark is here - https://github.com/VikParuchuri/marker/tree/master/benchmark... . Will run a full benchmark soon.

Mistral OCR is an impressive model, but OCR is a hard problem, and there is a significant risk of hallucinations/missing text with LLMs.

  • lolinder 3 days ago

    > with LLM as a judge

    For anyone else interested, prompt is here [0]. The model used was gemini-2.0-flash-001.

    Benchmarks are hard, and I understand the appeal of having something that seems vaguely deterministic rather than having a human in the loop, but I have a very hard time accepting any LLM-judged benchmarks at face value. This is doubly true when we're talking about something like OCR which, as you say, is a very hard problem for computers of any sort.

    I'm assuming you've given this some thought—how did you arrive at using an LLM to benchmark OCR vs other LLMs? What limitations with your benchmark have you seen/are you aware of?

    [0] https://github.com/VikParuchuri/marker/blob/master/benchmark...

    • themanmaran 2 days ago

      We also ran an OCR benchmark with LLM as judge using structured outputs. You can check out the full methodology on the repo [1]. But the general idea is:

      - Every document has ground truth text, a JSON schema, and the ground truth JSON.

      - Run OCR on each document and pass the result to GPT-4o along with the JSON Schema

      - Compare the predicted JSON against the ground truth JSON for accuracy.

      In our benchmark, the ground truth text => gpt-4o was 99.7%+ accuracy. Meaning whenever gpt-4o was given the correct text, it could extract the structured JSON values ~100% of the time. So if we pass in the OCR text from Mistral and it scores 70%, that means the inaccuracies are isolated to OCR errors.

      https://github.com/getomni-ai/benchmark

      • cdolan 2 days ago

        were you guys able to finish running the benchmark with mistral and got a 70% score? Missed that

        Edit - I see it on the Benchmark page now. Woof, low 70% scores in some areas!

        https://getomni.ai/ocr-benchmark

        • themanmaran 2 days ago

          Yup, surprising results! We were able to dig in a bit more. Main culprit is the overzealous "image extraction". Where if Mistral classifies something as an image, it will replace the entire section with (image)[image_002).

          And it happened with a lot of full documents as well. Ex: most receipts got classified as images, and so it didn't extract any text.

          • cdolan 2 days ago

            This sounds like a real problem and hurdle for North American (US/CAN in particular) invoice and receipt processing?

          • lingjiekong 2 days ago

            where do you find this regarding "Where if Mistral classifies something as an image, it will replace the entire section with (image)[image_002)."?

            • culi 2 days ago

              themanmaran works at Omni so presumably they have access to the actual resulting data from this study

      • someothherguyy 2 days ago

        Wouldn't that just bias itself to the shape of the text extracted from the OCR against the shape of the raw text alone? It doesn't seem like it would be a great benchmark for estimating semantic accuracy?

    • vikp 3 days ago

      Benchmarking is hard for markdown because of the slight formatting variations between different providers. With HTML, you can use something like TEDS (although there are issues with this, too), but with markdown, you don't have a great notion of structure, so you're left with edit distance.

      I think blockwise edit distance is better than full page (find the ground truth blocks, then infer each block separately and compare), but many providers only do well on full pages, which doesn't make it fair.

      There are a few different benchmark types in the marker repo:

        - Heuristic (edit distance by block with an ordering score)
        - LLM judging against a rubric
        - LLM win rate (compare two samples from different providers)
      
      None of these are perfect, but LLM against a rubric has matched visual inspection the best so far.

      I'll continue to iterate on the benchmarks. It may be possible to do a TEDS-like metric for markdown. Training a model on the output and then benchmarking could also be interesting, but it gets away from measuring pure extraction quality (the model benchmarking better is only somewhat correlated with better parse quality). I haven't seen any great benchmarking of markdown quality, even at research labs - it's an open problem.

    • arthurcolle 3 days ago

      You can use structured outputs, or something like my https://arthurcolle--dynamic-schema.modal.run/

      to extract real data from unstructured text (like that producted from an LLM) to make benchmarks slightly easier if you have a schema

      • cdolan 2 days ago

        What is the project? It just returns a vanilla html page saying:

        Dynamic Schema API API is running. See documentation for available endpoints.

  • carlgreene 3 days ago

    Thank you for your work on Marker. It is the best OCR for PDFs I’ve found. The markdown conversion can get wonky with tables, but it still does better than anything else I’ve tried

    • vikp 3 days ago

      Thanks for sharing! I'm training some models now that will hopefully improve this and more :)

  • netdevphoenix 2 days ago

    LLM as a judge?

    Isn't that a potential issue? You are assuming the LLM judge is reliable. What evidence do you have to assure yourself or/and others that it is reasonable assumption

    • bfors 2 days ago

      Perhaps they already evaluated their LLM judge model (with another LLM)

  • ChrisRob 2 days ago

    Thanks for the tip. Marker solved a table conversion without LLM that docling wasn't able to solve.

  • ntkris 2 days ago

    This is awesome. Have you seen / heard of any benchmarks where the data is actually a structured JSON vs. markdown?

  • codelion 3 days ago

    Really interesting benchmark, thanks for sharing! It's good to see some real-world comparisons. The hallucinations issue is definitely a key concern with LLM-based OCR, and it's important to quantify that risk. Looking forward to seeing the full benchmark results.

  • stavros 2 days ago

    I like the licensing options! Hopefully they make enough money to fund development.

  • DeathArrow 2 days ago

    >Mistral OCR is an impressive model, but OCR is a hard problem, and there is a significant risk of hallucinations/missing text with LLMs.

    To fight hallucinations, can't we use more LLMs and pick blocks where the majority of LLMs agree?

    • boxed 2 days ago

      Why wouldn't hallucinations be agreed upon if they have roughly the same training data?

      • TJSomething 2 days ago

        A hallucination is often an indication that the model doesn't know something. Then, the internal signal gets dominated by noise from the seeded training weights. Efforts to eliminate hallucinations with a single model have found success by asking the same question in different ways and only taking answers that agree. Logically, you could get more durable results from multiple models on the same prompt.

        • boxed 2 days ago

          I don't see why any of that makes logical sense. These models require such enormous training data that they pretty much MUST use the same training data to a very large degree. The training data itself is what they spit out. So "hallucinations" are just the training data you get out, which is the entire point of the models in the first place. There is no difference between an hallucination and a correct answer from the perspective of the math.

          • neuronic 2 days ago

            Isn' it just statistical word pattern prediction based on training data? These models likely don't "know" something anyway and cannot verify "truth" and facts. Reasoning attempts seem to me basically just like looping until the model finds a self-satisfying equilibrium state with different output.

            In that way, LLMs are more human than, say, a database or a book containing agreed-upon factual information which can be directly queried on demand.

            Imagine if there was just ONE human with human limitations on the entire planet who was taught everything for a long time - how reliable do you think they are with information retrieval? Even highly trained individuals (e.g. professors) can get stuff wrong on their specific topics at times. But this is not what we expect and demand from computers.

bambax 3 days ago

It's not bad! But it still hallucinates. Here's an example of an (admittedly difficult) image:

https://i.imgur.com/jcwW5AG.jpeg

For the blocks in the center, it outputs:

> Claude, duc de Saint-Simon, pair et chevalier des ordres, gouverneur de Blaye, Senlis, etc., né le 16 août 1607 , 3 mai 1693 ; ép. 1○, le 26 septembre 1644, Diane - Henriette de Budos de Portes, morte le 2 décembre 1670; 2○, le 17 octobre 1672, Charlotte de l'Aubespine, morte le 6 octobre 1725.

This is perfect! But then the next one:

> Louis, commandeur de Malte, Louis de Fay Laurent bre 1644, Diane - Henriette de Budos de Portes, de Cressonsac. du Chastelet, mortilhomme aux gardes, 2 juin 1679.

This is really bad because

1/ a portion of the text of the previous bloc is repeated

2/ a portion of the next bloc is imported here where it shouldn't be ("Cressonsac"), and of the right most bloc ("Chastelet")

3/ but worst of all, a whole word is invented, "mortilhomme" that appears nowhere in the original. (The word doesn't exist in French so in that case it would be easier to spot; but the risk is when words are invented, that do exist and "feel right" in the context.)

(Correct text for the second bloc should be:

> Louis, commandeur de Malte, capitaine aux gardes, 2 juin 1679.)

  • layer8 3 days ago

    > This is perfect!

    Just a nit, but I wouldn’t call it perfect when using U+25CB ○ WHITE CIRCLE instead of what should be U+00BA º MASCULINE ORDINAL INDICATOR, or alternatively a superscript “o”. These are https://fr.wikipedia.org/wiki/Adverbe_ordinal#Premiers_adver....

    There’s also extra spaces after the “1607” and around the hyphen in “Diane-Henriette”.

    Lastly, U+2019 instead of U+0027 would be more appropriate for the apostrophe, all the more since in the image it looks like the former and not like the latter.

    • MatthiasPortzel 3 days ago

      Slightly unrelated, but I once used Apple’s built-in OCR feature LiveText to copy a short string out of an image. It appeared to work, but I later realized it had copied “M” as U+041C (Cyrillic Capital Letter Em), causing a regex to fail to match. OCR giving identical characters is only good enough until it’s not.

    • jorvi 3 days ago

      > Just a nit, but I wouldn’t call it perfect when using U+25CB ○ WHITE CIRCLE instead of what should be U+00BA º MASCULINE ORDINAL INDICATOR, or alternatively a superscript “o”

      Or degree symbol. Although it should be able to figure out which to use according to the context.

    • TeMPOraL 3 days ago

      This is "reasoning model" stuff even for humans :).

      • layer8 3 days ago

        There is OCR software that analyses which language is used, and then applies heuristics for the recognized language to steer the character recognition in terms of character sequence likelihoods and punctuation rules.

        I don’t think you need a reasoning model for that, just better training; although conversely a reasoning model should hopefully notice the errors — though LLM tokenization might still throw a wrench into that.

    • raffraffraff 2 days ago

      It feels like, after the OCR step there should be language and subject matter detection, with a final sweep with a spelling / grammar checker that has the right "dictionary" selected. (That, right there, is my naivety on the subject, but I would have thought that the type of problem you're describing isn't OCR but classical spelling and grammar checking?)

      • layer8 2 days ago

        It’s OCR because the wrong characters are being recognized. This is not about fixing spelling or punctuation mistakes present in the source image, it’s that errors are being introduced, due to a lack of accuracy of this OCR with regard to punctuation and typography. The punctuation errors are not different in principle from the case of the OCR producing a misspelled word that wasn’t misspelled in the image being OCRed.

        A subsequent cleanup pass that fixes grammar/spelling errors, as you propose, wouldn’t be appropriate when the goal is to faithfully reproduce the original text.

        And specifically for the “white circle” character, it would be difficult to correctly infer the original ordinal markers after the fact. I myself could only do so by inspecting the original image, i.e. by having my brain redo the OCR.

        • raffraffraff 2 days ago

          > A subsequent cleanup pass that fixes grammar/spelling errors, as you propose, wouldn’t be appropriate when the goal is to faithfully reproduce the original text

          I suppose that depends on why it's wrong. Did the model accurately read a real typo in the image or did it incorrectly decipher a character? If a spelling & grammar pass fixes the latter, isn't it valid?

      • pbhjpbhj 2 days ago

        Not unrelated - OneNote 'copy text from image' has started producing lots of incorrect OCR results, but they're all non-words.

        For example, from a clear image of a printed page (in a standard font), it will give me 'cornprising' instead of 'comprising'; 'niatter' instead of 'matter'. Excepting the spell-check underline they'd be hard to spot as with relatively tight kerning all the errors look like the originals.

        I'm surprised as 1) I've not had these sorts of errors before, 2) they're not words, and words must be heavily weighted for in the OCR engine (I'd have thought).

  • bambax 3 days ago

    Another test with a text in English, which is maybe more fair (although Mistral is a French company ;-). This image is from Parliamentary debates of the parliament of New Zealand in 1854-55:

    https://i.imgur.com/1uVAWx9.png

    Here's the output of the first paragraph, with mistakes in brackets:

    > drafts would be laid on the table, and a long discussion would ensue; whereas a Committee would be able to frame a document which, with perhaps a few verbal emundations [emendations], would be adopted; the time of the House would thus be saved, and its business expected [expedited]. With regard to the question of the comparative advantages of The-day [Tuesday]* and Friday, he should vote for the amendment, on the principle that the wishes of members from a distance should be considered on all sensations [occasions] where a principle would not be compromised or the convenience of the House interfered with. He hoped the honourable member for the Town of Christchurch would adopt the suggestion he (Mr. Forssith [Forsaith]) had thrown out and said [add] to his motion the names of a Committee.*

    Some mistakes are minor (emnundations/emendations or Forssith/Forsaith), but others are very bad, because they are unpredictable and don't correspond to any pattern, and therefore can be very hard to spot: sensations instead of occasions, or expected in lieu of expedited... That last one really changes the meaning of the sentence.

  • spudlyo 3 days ago

    I want to rejoice that OCR is now a "solved" problem, but I feel like hallucinations are just as problematic as the kind of stuff I have to put up with tesseract -- both require careful manual proofreading for an acceptable degree of confidence. I guess I'll have to try it and see for myself just how much better these solutions are for my public domain archive.org Latin language reader & textbook projects.

    • qingcharles 3 days ago

      It depends on your use-case. For mine, I'm mining millions of scanned PDF pages to get approximate short summaries of long documents. The occasional hallucination won't damage the project. I realize I'm an outlier, and I would obviously prefer a solution that was as accurate as possible.

    • eMPee584 3 days ago

      possibly doing both & diffing the output to spot contested bits?

      • spudlyo 3 days ago

        that’s my current idea, use two different ocr models and diff the results to spot check for errors. at these prices why not?

  • thomasfromcdnjs 3 days ago

    Does anyone know the correlation between our abilities to parse PDF's and the quality of our LLM's training datasets?

    If a lot of scientific papers have been pdf's and hitherto had bad conversions to text/tokens, can we expect to see major gains in our training and therefore better outputs?

  • rossant 2 days ago

    Your example doesn't seem that difficult to me.

  • Kokichi 3 days ago

    All it ever does is hallucinate

owenpalmer 3 days ago

This is incredibly exciting. I've been pondering/experimenting on a hobby project that makes reading papers and textbooks easier and more effective. Unfortunately the OCR and figure extraction technology just wasn't there yet. This is a game changer.

Specifically, this allows you to associate figure references with the actual figure, which would allow me to build a UI that solves the annoying problem of looking for a referenced figure on another page, which breaks up the flow of reading.

It also allows a clean conversion to HTML, so you can add cool functionality like clicking on unfamiliar words for definitions, or inserting LLM generated checkpoint questions to verify understanding. I would like to see if I can automatically integrate Andy Matuschak's Orbit[0] SRS into any PDF.

Lots of potential here.

[0] https://docs.withorbit.com/

  • NalNezumi 3 days ago

    >a UI that solves the annoying problem of looking for a referenced figure on another page, which breaks up the flow of reading.

    A tangent but this exact issue is what I was frustrated for a long time with pdf reader and reading science papers. Then I found sioyek that pops up a small window when you hover over links (references and equations and figures) and it solved it.

    Granted, the pdf file must be in right format, so OCR could make this experience better. Just saying the UI component of that already exist

    https://sioyek.info/

    • PerryStyle 3 days ago

      Zotero's PDF viewer also does this now. Being able to annotate PDFs and having a reference manager has been a life saver.

    • owenpalmer 3 days ago

      Thanks for the link! Good to know someone is working on something similar.

  • generalizations 3 days ago

    Wait does this deal with images?

    • ezfe 3 days ago

      The output includes images from the input. You can see that on one of the examples where a logo is cropped out of the source and included in the result.

raunakchowdhuri 3 days ago

We ran some benchmarks comparing against Gemini Flash 2.0. You can find the full writeup here: https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini

A high level summary is that while this is an impressive model, it underperforms even current SOTA VLMs on document parsing and has a tendency to hallucinate with OCR, table structure, and drop content.

  • shrisukhani 2 days ago

    Anecdotally, we also found Gemini Flash to be better.

  • hackernewds 2 days ago

    meanwhile, you're comparing it to the output of almost a trillion dollar company

    • stann 2 days ago

      The tagline boasts that it is "introducing the world’s best document understanding API". So, holding them to their marketing seems fair

      • neuronic 2 days ago

        Isn't anyone who releases anything putting "the world's best blablabla" on their page nowadays? I've become entirely blind to it.

        • dwedge 2 days ago

          If they put it, and it's subpar, I write off the product.

    • HaZeust 2 days ago

      ... And? We're judging it for the merits of the technology it purports to be, not the pockets of the people that bankroll them. Probably not fair - sure, but when I pick my OCR, I want to pick SOTA. These comparisons and announcements help me find those.

Asraelite 3 days ago

I never thought I'd see the day where technology finally advanced far enough that we can edit a PDF.

  • randomNumber7 3 days ago

    I never thought driving a car is harder than editing a pdf.

    • pzo 3 days ago

      It's not about harder but about what error you can tolerate. Here if you have accuracy 99% for many applications it's enough. If you have 99% accuracy per trip of no crash during self driving then you gonna be dead within a year very likely.

      For cars we need accuracy at least 99.99% and that's very hard.

      • rtsil 3 days ago

        I doubt most people have 99% accuracy. The threshold of tolerance for error is just much lower for any self-driving system (and with good reason, because we're not familiar with them yet).

        • KeplerBoy 3 days ago

          How do you define 99% accuracy?

          I guess something like success rate for a trip (or mile) would be a more reasonable metric. Most people have a success rate far higher than 99% for averages trips.

          Most people who commute daily are probably doing something like a 1000 car rides a year and have minor accidents every few years. 99% success rates would mean monthly accidents.

  • toephu2 3 days ago

    I've been able to edit PDFs (95%+ of them) accurately for the past 10 years...

  • Apofis 3 days ago

    Foxit PDF exists...

kbyatnal 3 days ago

We're approaching the point where OCR becomes "solved" — very exciting! Any legacy vendors providing pure OCR are going to get steamrolled by these VLMs.

However IMO, there's still a large gap for businesses in going from raw OCR outputs —> document processing deployed in prod for mission-critical use cases. LLMs and VLMs aren't magic, and anyone who goes in expecting 100% automation is in for a surprise.

You still need to build and label datasets, orchestrate pipelines (classify -> split -> extract), detect uncertainty and correct with human-in-the-loop, fine-tune, and a lot more. You can certainly get close to full automation over time, but it's going to take time and effort. But the future is on the horizon!

Disclaimer: I started a LLM doc processing company to help companies solve problems in this space (https://extend.app/)

  • dml2135 3 days ago

    One problem I’ve encountered at my small startup in evaluating OCR technologies is precisely convincing stakeholders that the “human-in-the-loop” part is both unavoidable, and ultimately beneficial.

    PMs want to hear that an OCR solution will be fully automated out-of-the-box. My gut says that anything offering that is snake-oil, and I try to convey that the OCR solution they want is possible, but if you are unwilling to pay the tuning cost, it’s going to flop out of the gate. At that point they lose interest and move on to other priorities.

    • kbyatnal 3 days ago

      Yup definitely, and this is exactly why I built my startup. I've heard this a bunch across startups & large enterprises that we work with. 100% automation is an impossible target, because even humans are not 100% perfect. So how we can expect LLMs to be?

      But that doesn't mean you have to abandon the effort. You can still definitely achieve production-grade accuracy! It just requires having the right tooling in place, which reduces the upfront tuning cost. We typically see folks get there on the order of days or 1-2 weeks (it doesn't necessarily need to take months).

    • golergka 3 days ago

      It really depends on their fault tolerance. I think there's a ton of useful applications where OCR would be 99.9%, 99%, and even 98% reliable. Skillful product manager can keep these limitations in mind and work around them.

    • jocoda 2 days ago

      ... unavoidable "human in the loop" - depends imo.

      From the comments here, it certainly seems that for general OCR it's not up to snuff yet. Luckily, I don't have great ambitions.

      I can see this working for me with just a little care upfront preprocessing now that I know where it falls over. It casually skips portions of the document, and misses certain lines consistently. Knowing that I can do a bit massaging, and feed it what I know it likes, and then reassemble.

      I found in testing that it failed consistently at certain parts, but where it worked, it worked extremely well in contrast to other methods/services that I've been using.

  • risyachka 3 days ago

    >> Any legacy vendors providing pure OCR are going to get steamrolled by these VLMs.

    -OR- they can just use these APIs, and considering that they have a client base - which would prefer to not rewrite integrations to get the same result - they can get rid of most code base, replace it with llm api and increase margins by 90% and enjoy good life.

    • esafak 3 days ago

      They're going to become commoditized unless they add value elsewhere. Good news for customers.

      • TeMPOraL 3 days ago

        They are (or at least could easily be) adding value in form of SLA - charging money for giving guarantees on accuracy. This is both better for customer, who gets concrete guarantees and someone to shift liability to, and for the vendor, that can focus on creating techniques and systems for getting that extra % of reliability out of the LLM OCR process.

        All of the above are things companies - particularly larger ones - are happy to pay for, because ORC is just a cog in the machine, and this makes it more reliable and predictable.

        On top of the above, there are auxiliary value-adds such a vendor could provide - such as, being fully compliant with every EU directive and regulation that's in power, or about to be. There's plenty of those, they overlap, and no one wants to deal with it if they can outsource it to someone who already figured it out.

        (And, again, will take the blame for fuckups. Being a liability sink is always a huge value-add, in any industry.)

  • techwizrd 3 days ago

    The challenge I have is how to get bounding boxes for the OCR, for things like redaction/de-identification.

    • dontlikeyoueith 3 days ago

      AWS Textract works pretty well for this and is much cheaper than running LLMs.

      • daemonologist 3 days ago

        Textract is more expensive than this (for your first 1M pages per month at least) and significantly more than something like Gemini Flash. I agree it works pretty well though - definitely better than any of the open source pure OCR solutions I've tried.

    • kbyatnal 3 days ago

      yeah that's a fun challenge — what we've seen work well is a system that forces the LLM to generate citations for all extracted data, map that back to the original OCR content, and then generate bounding boxes that way. Tons of edge cases for sure that we've built a suite of heuristics for over time, but overall works really well.

  • einpoklum 3 days ago

    An LLM with billions of parameters for extracting text from a PDF (which isn't even a rasterized image) really does not "solve OCR".

  • nextworddev 3 days ago

    Your customer includes Checkr? Impressive. Are they referencable?

    • wahnfrieden a day ago

      btw - what 'dark patterns' does portkey contain?

mvac 3 days ago

Great progress, but unfortunately, for our use case (converting medical textbooks from PDF to MD), the results are not as good as those by MinerU/PDF-Extract-Kit [1].

Also the collab link in the article is broken, found a functional one [2] in the docs.

[1] https://github.com/opendatalab/MinerU [2] https://colab.research.google.com/github/mistralai/cookbook/...

  • owenpalmer 3 days ago

    I've been searching relentlessly for something like this! I wonder why it's been so hard to find... is it the Chinese?

    In any case, thanks for sharing.

  • thelittleone 3 days ago

    Have you had a chance to compare results from MinerU vs LLM such a Gemini 2.0 or anthropic's native PDF tool?

    • mvac 2 days ago

      Yes, i have. The problem with using just an LLM is that while it reads and understands text, but it cannot reproduce it accurately. Additionaly the textbooks I've mentioned have many diagrams and illustrations in them (e.g. books on anatomy or biochemistry). I don't really care about extracting text from them, I just need them extracted as images alongside the text, and no LLM does that.

shekhargulati 3 days ago

Mistral OCR made multiple mistakes in extracting this [1] document. It is a two-page-long PDF in Arabic from the Saudi Central Bank. The following errors were observed:

- Referenced Vision 2030 as Vision 2.0. - Failed to extract the table; instead, it hallucinated and extracted the text in a different format. - Failed to extract the number and date of the circular.

I tested the same document with ChatGPT, Claude, Grok, and Gemini. Only Claude 3.7 extracted the complete document, while all others failed badly. You can read my analysis here [2].

1. https://rulebook.sama.gov.sa/sites/default/files/en_net_file... 2. https://shekhargulati.com/2025/03/05/claude-3-7-sonnet-is-go...

vessenes 3 days ago

Dang. Super fast and significantly more accurate than google, Claude and others.

Pricing : $1/1000 pages, or per 2k pages if “batched”. I’m not sure what batching means in this case: multiple pdfs? Why not split them to halve the cost?

Anyway this looks great at pdf to markdown.

  • sophiebits 3 days ago

    Batched often means a higher latency option (minutes/hours instead of seconds), which providers can schedule more efficiently on their GPUs.

  • abiraja 3 days ago

    Batching likely means the response is not real-time. You set up a batch job and they send you the results later.

    • ozim 3 days ago

      If only business people I work with would understand 100GB even transfer over the network is not going to return immediately results ;)

    • vessenes 3 days ago

      That makes sense. Idle time is nearly free after all.

  • kapitalx 3 days ago

    From my testing so far, it seems it's super fast and responded synchronously. But it decided that the entire page is an image and returned `![img-0.jpeg](img-0.jpeg)` with coordinates in the metadata for the image, which is the entire page.

    Our tool, doctly.ai is much slower and async, but much more accurate and gets you the content itself as an markdown.

    • ralusek 3 days ago

      I thought we stopped -ly company names ~8 years ago?

      • kapitalx 3 days ago

        Haha for sure. Naming isn't just the hardest problem in computer science, it's always hard. But at some point you just have to pick something and move forward.

      • yieldcrv 3 days ago

        if you talk to people gen-x and older, you still need .com domains

        for all those people that aren't just clicking on a link on their social media feed, chat group, or targeted ad

  • Tostino 3 days ago

    Usually (With OpenAI, I haven't checked Mistral yet) it means an async api rather than a sync api.

    e.g. you submit multiple requests (pdfs) in one call, and get back an id for the batch. You then can check on the status of that batch and get the results for everything when done.

    It lets them use their available hardware to it's full capacity much better.

  • odiroot 3 days ago

    May I ask as a layperson, how would you about using this to OCR multiple hundreds of pages? I tried the chat but it pretty much stops after the 2nd page.

    • beklein 3 days ago

      You can check the example code on the Mistral documentation, you would _only_ have to change the value of the variable `document_url` to the URL of your uploaded PDF... and you need to change the `MISTRAL_API_KEY` to the value of your specific key that you can get from the Le Platforme webpage.

      https://docs.mistral.ai/capabilities/document/#ocr-with-pdf

    • sneak 3 days ago

      Submit the pages via the API.

      • odiroot 3 days ago

        This worked indeed. Although I had to cut my document into smaller chunks. 900 pages at once ended with a timeout.

  • jacksnipe 3 days ago

    I would assume this is 1 request containing 2k pages vs N requests whose total pages add up to 1000.

serjester 3 days ago

This is cool! With that said for anyone looking to use this in RAG, the downside to specialized models instead of general VLMs is you can't easily tune it to your use specific case. So for example, we use Gemini to add very specific alt text to images in the extracted Markdown. It's also 2 - 3X the cost of Gemini Flash - hopefully the increased performance is significant.

Regardless excited to see more and more competition in the space.

Wrote an article on it: https://www.sergey.fyi/articles/gemini-flash-2-tips

  • hyuuu 3 days ago

    gemini flash is notorious for hallucinating the output of the OCR, be careful with it. For straight forward, semi-structured, low page count (under 5) it should perform well, but the more the context window is stretched the more the output gets more unreliable

sbarre 3 days ago

6 years ago I was working with a very large enterprise that was struggling to solve this problem, trying to scan millions of arbitrary forms and documents per month to clearly understand key points like account numbers, names and addresses, policy numbers, phone numbers, embedded images or scribbled notes, and also draw relationships between these values on a given form, or even across forms.

I wasn't there to solve that specific problem but it was connected to what we were doing so it was fascinating to hear that team talk through all the things they'd tried, from brute-force training on templates (didn't scale as they had too many kinds of forms) to every vendor solution under the sun (none worked quite as advertised on their data)..

I have to imagine this is a problem shared by so many companies.

opwieurposiu 3 days ago

Related, does anyone know of an app that can read gauges from an image and log the number to influx? I have a solar power meter in my crawlspace, it is inconvenient to go down there. I want to point an old phone at it and log it so I can check it easily. The gauge is digital and looks like this:

https://www.pvh2o.com/solarShed/firstPower.jpg

  • ubergeek42 3 days ago

    This[1] is something I've come across but not had a chance to play with, designed for reading non-smart meters that might work for you. I'm not sure if there's any way to run it on an old phone though.

    [1] https://github.com/jomjol/AI-on-the-edge-device

    • jasonjayr 3 days ago

      Wow. I was looking at hooking my water meter into home assistant, and was going to investigate just counting an optical pulse (it has a white portion on the gear that is in a certain spot every .1 gal) This is like the same meter I use, and perfect.

      (It turns out my electric meter, though analog, blasts out it's reading on RF every 10 seconds unencrypted. I got that via my RTL-SDR reciever :) )

    • timc3 3 days ago

      I use this for a watermeter. Works quite well as long as you have a good SD card

  • dehrmann 3 days ago

    You'll be happier finding a replacement meter that has an interface to monitor it directly or a second meter. An old phone and OCR will be very brittle.

    • haswell 3 days ago

      Not OP, but it sounds like the kind of project I’d undertake.

      Happiness for me is about exploring the problem within constraints and the satisfaction of building the solution. Brittleness is often of less concern than the fun factor.

      And some kinds of brittleness can be managed/solved, which adds to the fun.

      • arcfour 3 days ago

        I would posit that learning how the device works, and how to integrate with a newer digital monitoring device would be just as interesting and less brittle.

        • haswell 3 days ago

          Possibly! But I’ve recently wanted to dabble with computer vision, so I’d be looking at a project like this as a way to scratch a specific itch. Again, not OP so I don’t know what their priorities are, but just offering one angle for why one might choose a less “optimal” approach.

  • ramses0 3 days ago

    https://www.home-assistant.io/integrations/seven_segments/

    https://www.unix-ag.uni-kl.de/~auerswal/ssocr/

    https://github.com/tesseract-ocr/tesseract

    https://community.home-assistant.io/t/ocr-on-camera-image-fo...

    https://www.google.com/search?q=home+assistant+ocr+integrati...

    https://www.google.com/search?q=esphome+ocr+sensor

    https://hackaday.com/2021/02/07/an-esp-will-read-your-meter-...

    ...start digging around and you'll likely find something. HA has integrations which can support writing to InfluxDB (local for sure, and you can probably configure it for a remote influxdb).

    You're looking at 1xRaspberry PI, 1xUSB Webcam, 1x"Power Management / humidity management / waterproof electrical box" to stuff it into, and then either YOLO and DIY to shoot over to your influxdb, or set up a Home Assistant and "attach" your frankenbox as some sort of "sensor" or "integration" which spits out metrics and yadayada...

  • renewiltord 3 days ago

    4o transcribes it perfectly. You can usually root an old Android and write this app in ~2h with LLMs if unfamiliar. The hard part will be maintaining camera lens cleanliness and alignment etc.

    The time cost is so low that you should give it a gander. You'll be surprised how fast you can do it. If you just take screenshots every minute it should suffice.

    • pavl 3 days ago

      What software-tools do you usw to Programm the APP?

      • renewiltord 3 days ago

        Since it's at home, you'll have WiFi access, so it's pretty much a rudimentary Kotlin app on Android. You can just grab a photo and ship it to the GPT-4o API, get the response, and then POST it somewhere.

  • BonoboIO 3 days ago

    Gemini Free Tier would surely work

evmar 3 days ago

I noticed on the Arabic example they lost a space after the first letter on the third to last line, can any native speakers confirm? (I only know enough Arabic to ask dumb questions like this, curious to learn more.)

Edit: it looks like they also added a vowel mark not present in the input on the line immediately after.

Edit2: here's a picture of what I'm talking about, the before/after: https://ibb.co/v6xcPMHv

  • resiros 3 days ago

    Arabic speaker here. No, it's perfect.

    • evmar 3 days ago

      I am pretty sure it added a kasrah not present in the input on the 2nd to last line. (Not saying it's not super impressive, and also that almost certainly is the right word, but I think that still means not quite "perfect"?)

      • gl-prod 3 days ago

        Yes, it looks like it did add a kasrah to the word ظهري

        • yoda97 3 days ago

          Yep, and فمِنا too, this is not just OCR, it made some post-processing corrections or "enhancements". That could be good, but it could also be trouble the 1% chance it makes a mistake in critical documents.

    • gl-prod 3 days ago

      He means the space between the wāw (و) and the word

      • evmar 3 days ago

        I added a pic to the original comment, sorry for not being clear!

  • albatrosstrophy 3 days ago

    And here I thought after reading the headline: finally a reliable Arabic OCR. I've never in my life found a good that does the job decently especially for a scanned document. Or is there something out there I don't know about?

neom 3 days ago

I gave it a bunch of my wifes 18th century English scans to transcribe, mostly couldn't do it, and it's been doing this for 15 minutes now, not sure why but i find quite amusing: https://share.zight.com/L1u2jZYl

ChemSpider 3 days ago

"World's best OCR model" - that is quite a statement. Are there any well-known benchmarks for OCR software?

  • themanmaran 3 days ago

    We published this benchmark the other week. We'll can update and run with Mistral today!

    https://github.com/getomni-ai/benchmark

    • themanmaran 3 days ago

      Update: Just ran our benchmark on the Mistral model and results are.. surprisingly bad?

      Mistral OCR:

      - 72.2% accuracy

      - $1/1000 pages

      - 5.42s / page

      Which is pretty far cry from the 95% accuracy they were advertising from their private benchmark. The biggest thing I noticed is how it skips anything it classifies as an image/figure. So charts, infographics, some tables, etc. all get lifted out and returned as [image](image_002). Compared to the other VLMs that are able to interpret those images into a text representation.

      https://github.com/getomni-ai/benchmark

      https://huggingface.co/datasets/getomni-ai/ocr-benchmark

      https://getomni.ai/ocr-benchmark

      • Thaxll 3 days ago

        Do you benchmark the right thing though? It seems to focus a lot on image / charts etc...

        The 95% from their benchmark: "we evaluate them on our internal “text-only” test-set containing various publication papers, and PDFs from the web; below:"

        Text only.

        • themanmaran 2 days ago

          Our goal is to benchmark on real world data. Which is often more complex than plain text. If we have to make the benchmark data easier for the model to perform better, it's not an honest assessment of the reality.

    • kergonath 3 days ago

      Excellent. I am looking forward to it.

    • cdolan 3 days ago

      Came here to see if you all had run a benchmark on it yet :)

  • WhitneyLand 3 days ago

    It’s interesting that none of the existing models can decode a Scrabble board screen shot and give an accurate grid of characters.

    I realize it’s not a common business case, came across it testing how well LLMs can solve simple games. On a side note, if you bypass OCR and give models a text layout of a board standard LLMs cannot solve Scrabble boards but the thinking models usually can.

  • resource_waste 3 days ago

    Its Mistral, they are the only homegrown AI Europe has, so people pretend they are meaningful.

    I'll give it a try, but I'm not holding my breath. I'm a huge AI Enthusiast and I've yet to be impressed with anything they've put out.

raffraffraff 2 days ago

Forgive my absolute ignorance, I should probably run this through a chat bot before posting ... So I'm updating my post with answers now!

Q: Do LLMs specialise in "document level" recognition based on headings, paragraphs, columns tables etc? Ie: ignore words and characters for now and attempt to recognise a known document format.

A: Not most LLMs, but those with multimodal / vision capability could (eg DeepSeek Vision. ChatGPT 4). There are specialized models for this work like Tesseract, LayoutLM.

Q: How did OCR work "back in the day" before we had these LLMs? Are any of these methods useful now?

A: They used pattern recognition and feature extraction, rules and templates. Newer ML based OCR used SVM to isolate individual characters and HMM to predict the next character or word. Today's multimodal models process images and words, can handle context better than the older methods, and can recognise whole words or phrases instead of having to read each character perfectly. This is why they can produce better results but with hallucinations.

Q: Can LLMs rate their own confidence in each section, maybe outputting text with annotations that say "only 10% certain of this word", and pass the surrounding block through more filters, different LLMs, different methods to try to improve that confidence?

A: Short answer, "no". But you can try to estimate with post processing.

Or am I super naive, and all of those methods are already used by the big commercial OCR services like Textract etc?

SilentM68 3 days ago

I would like to see how it performs with massively warped and skewed scanned text images, basically a scanned image where the text lines are wavy as opposed as straight horizontal, where the letters are elongated. One where the line widths are different depending on the position on the scanned image. I once had to deal with such a task that somebody gave me with OCR software, Acrobat, and other tools could not decode the mess so I had to recreate the 30 pages myself, manually. Not a fun thing to do but that is a real use case.

  • thegabriele 2 days ago

    I use gemini to solve textual CAPTCHAS with those kind of distortions and more: 60% of the time it works every time.

  • amelius 3 days ago

    Are you trying to build a captcha solver?

    • SilentM68 3 days ago

      No, not a captcha solver. When I worked in education, I was given a 90s paper document that a teacher needed OCRd but it was completely warped. It was my job to remediate those type of documents for Accessibility reasons. I had to scan and OCR it but the result was garbage. Mind you I had access to Windows, Linux and MacOS tools but still difficult to do. I had to guess what it said, which was not impossible but it was time-consuming, not doable in the time-frame I was given, so I had no option but to manually retype all the information into a new document and convert it that way. Document remediation and accessibility should be a good use case for A.I., in education.

  • arcfour 3 days ago

    Garbage in, garbage out?

    • edude03 3 days ago

      "Yes" but if a human could do it "AI" should be able to do it too.

sireat 2 days ago

Intriguing announcement, however the examples on the mistral.ai page seem rather "easy".

What about rare glyphs in different languages using handwriting from previous centuries?

I've been dealing with OCR issues and evaluating different approaches for past 5+ years at a national library that I work at.

Usual consensus is that widely used open source Tesseract is subpar to commercial models.

That might be so without fine tuning. However one can perform supplemental training and build your own Tesseract models that can outperform the base ones.

Case study of Kant's letter's from 18th century:

About 6 months ago, I tested OpenAi approach to OCR to some old 18th century letters that needed digitizing.

The results were rather good (90+% accuracy) with the usual hallucination here and there.

What was funny that OpenAI was using base Tesseract to generate the segmenting and initial OCR.

The actual OCRed content before last inference step was rather horrid because the Tesseract model that OpenAi was using was not appropriate for the particular image.

When I took OpenAi off the first step and moved to my own Tesseract models, I gained significantly in "raw" OCR accuracy at character level.

Then I performed normal LLM inference at the last step.

What was a bit shocking: My actual gains for the task (humanly readable text for general use) were not particularly significant.

That is LLMs are fantastic at "untangling" complete mess of tokens into something humanly readable.

For example:

P!3goattie -> prerogative (that is given the surrounding text is similarly garbled)

janalsncm 3 days ago

The hard ones are things like contracts, leases, and financial documents which 1) don’t have a common format 2) are filled with numbers proper nouns and addresses which it’s really important not to mess up 3) cannot be inferred from context.

Typical OCR pipeline would be to pass the doc through a character-level OCR system then correct errors with a statistical model like an LLM. An LLM can help correct “crodit card” to “credit card” but it cannot correct names or numbers. It’s really bad if it replaces a 7 with a 2.

cxie 3 days ago

The new Mistral OCR release looks impressive - 94.89% overall accuracy and significantly better multilingual support than competitors. As someone who's built document processing systems at scale, I'm curious about the real-world implications.

Has anyone tried this on specialized domains like medical or legal documents? The benchmarks are promising, but OCR has always faced challenges with domain-specific terminology and formatting.

Also interesting to see the pricing model ($1/1000 pages) in a landscape where many expected this functionality to eventually be bundled into base LLM offerings. This feels like a trend where previously encapsulated capabilities are being unbundled into specialized APIs with separate pricing.

I wonder if this is the beginning of the componentization of AI infrastructure - breaking monolithic models into specialized services that each do one thing extremely well.

  • themanmaran 3 days ago

    Excited to test this our on our side as well. We recently built an OCR benchmarking framework specifically for VLMs[1][2], so we'll do a test run today.

    From our last benchmark run, some of these numbers from Mistral seem a little bit optimistic. Side by side of a few models:

    model | omni | mistral |

    gemini | 86% | 89% |

    azure | 85% | 89% |

    gpt-4o | 75% | 89% |

    google | 68% | 83% |

    Currently adding the Mistral API and we'll get results out today!

    [1] https://github.com/getomni-ai/benchmark

    [2] https://huggingface.co/datasets/getomni-ai/ocr-benchmark

    • themanmaran 3 days ago

      Update: Just ran our benchmark on the Mistral model and results are.. surprisingly bad?

      Mistral OCR:

      - 72.2% accuracy

      - $1/1000 pages

      - 5.42s / page

      Which is pretty far cry from the 95% accuracy they were advertising from their private benchmark. The biggest thing I noticed is how it skips anything it classifies as an image/figure. So charts, infographics, some tables, etc. all get lifted out and returned as [image](image_002). Compared to the other VLMs that are able to interpret those images into a text representation.

      https://github.com/getomni-ai/benchmark

      https://huggingface.co/datasets/getomni-ai/ocr-benchmark

      https://getomni.ai/ocr-benchmark

    • jaggs 3 days ago

      By optimistic, do you mean 'tweaked'? :)

  • epolanski 3 days ago

    At my client we want to provide an AI that can retrieve relevant information from documentation (home building business, documents detail how to install a solar panel or a shower, etc) and we've set up an entire system with benchmarks, agents, etc, yet the bottleneck is OCR!

    We have millions and millions of pages of documents and an off by 1 % error means it compounds with the AI's own error, which compounds with documentation itself being incorrect at times, which leads it all to be not production ready (and indeed the project has never been released), not even close.

    We simply cannot afford to give our customers incorrect informatiin

    We have set up a backoffice app that when users have questions, it sends it to our workers along the response given by our AI application and the person can review it, and ideally correct the ocr output.

    Honestly after an year of working it feels like AI right now can only be useful when supervised all the time (such as when coding). Otherwise I just find LLMs still too unreliable besides basic bogus tasks.

    • PeterStuer 3 days ago

      As someone who has had a home built, and nearly all my friends and acquaintances report the same thing, having a 1% error on information in this business would mean not a 10x but a 50x improvement over the current practice in the field.

      If nobody is supervising building documents all the time during the process, every house would be a pile of rubbish. And even when you do stuff stills creeps in and has to be redone, often more than once.

  • janalsncm 3 days ago

    I have done OCR on leases. It’s hard. You have to be accurate and they all have bespoke formatting.

    It would almost be easier to switch everyone to a common format and spell out important entities (names, numbers) multiple times similar to how cheques do.

    The utility of the system really depends on the makeup of that last 5%. If problematic documents are consistently predictable, it’s possible to do a second pass with humans. But if they’re random, then you have to do every doc with humans and it doesn’t save you any time.

  • PeterStuer 3 days ago

    I'd love to try it for my domain (regulation), but $1/1000 pages is significantly more expensive than my current local Docling based setup that already does a great job of processing PDF's for my needs.

    • yawnxyz 3 days ago

      I think for regulated fields / high impact fields $1/1000 is well-worth the price; if the accuracy is close to 100% this is way better than using people, who are still error-prone

      • PeterStuer 2 days ago

        It could be very well worth the price, but it still needs to justify the price increase over an already locally running solution that is nearly free in operation.

        I will still check it out, but given the performance I already have for my specific use case with my current system, my upfront expectation is that it probably will not make it to production.

        I'm sure there are other applications for wich this could be a true enabler.

        I am also biased to using as little SaaS as possible. I prefer services on-prem and under my control where possible.

        I do use GPT-4o for now as, again, for my use case, it significantly outperformed other local solutions I tried.

  • kbyatnal 3 days ago

    re: real world implications, LLMs and VLMs aren't magi, and anyone who goes in expecting 100% automation is in for a surprise (especially in domains like medical or legal).

    IMO there's still a large gap for businesses in going from raw OCR outputs —> document processing deployed in prod for mission-critical use cases.

    e.g. you still need to build and label datasets, orchestrate pipelines (classify -> split -> extract), detect uncertainty and correct with human-in-the-loop, fine-tune, and a lot more. You can certainly get close to full automation over time, but it's going to take time and effort.

    But for RAG and other use cases where the error tolerance is higher, I do think these OCR models will get good enough to just solve that part of the problem.

    Disclaimer: I started a LLM doc processing company to help companies solve problems in this space (https://extend.app/)

  • kergonath 3 days ago

    > Has anyone tried this on specialized domains like medical or legal documents?

    I’ll try it on a whole bunch of scientific papers ASAP. Quite excited about this.

  • janis1234 3 days ago

    $1 for 1000 pages seems high to me. Doing a google search

    Rent and Reserve NVIDIA A100 GPU 80GB - Pricing Starts from $1.35/hour

    I just don't know if in 1 hour and with a A100 I can process more than a 1000 pages. I'm guessing yes.

    • blackoil 3 days ago

      Is the model Open Source/Weight? Else the cost is for the model, not GPU.

  • salynchnew 3 days ago

    Also interesting to see that parts of the training infrastructure to create frontier models is itself being monetized.

  • amelius 3 days ago

    > 94.89% overall accuracy

    There are about 47 characters on average in a sentence. So does this mean it gets around 2 or 3 mistakes per sentence?

  • stavros 3 days ago

    What do you mean by "free"? Using the OpenAI vision API, for example, for OCR is quite a bit more expensive than $1/1k pages.

  • unboxingelf 3 days ago

    We’ll just stick LLM Gateway LLM in front of all the specialized LLMs. MicroLLMs Architecture.

    • cxie 3 days ago

      I actually think you're onto something there. The "MicroLLMs Architecture" could mirror how microservices revolutionized web architecture.

      Instead of one massive model trying to do everything, you'd have specialized models for OCR, code generation, image understanding, etc. Then a "router LLM" would direct queries to the appropriate specialized model and synthesize responses.

      The efficiency gains could be substantial - why run a 1T parameter model when your query just needs a lightweight OCR specialist? You could dynamically load only what you need.

      The challenge would be in the communication protocol between models and managing the complexity. We'd need something like a "prompt bus" for inter-model communication with standardized inputs/outputs.

      Has anyone here started building infrastructure for this kind of model orchestration yet? This feels like it could be the Kubernetes moment for AI systems.

      • arcfour 3 days ago

        This is already done with agents. Some agents only have tools and the one model, some agents will orchestrate with other LLMs to handle more advanced use cases. It's pretty obvious solution when you think about how to get good performance out of a model on a complex task when useful context length is limited: just run multiple models with their own context and give them a supervisor model—just like how humans organize themselves in real life.

      • fnordpiglet 3 days ago

        I’m doing this personally for my own project - essentially building an agent graph that starts with the image output, orients and cleans, does a first pass with tesseract LSTM best models to create PDF/HOCR/Alto, then pass to other LLMs and models based on their strengths to further refine towards markdown and latex. My goal is less about RAG database population but about preserving in a non manually typeset form the structure and data and analysis, and there seems to be pretty limited tooling out there since the goal generally seems to be the obviously immediately commercial goal of producing RAG amenable forms that defer the “heavy” side of chart / graphic / tabular reproduction to a future time.

      • unboxingelf 3 days ago

        Take a look at MCP, Model Context Protocol.

notepad0x90 3 days ago

I was just watching a science-related video containing math equations. I wondered how soon will I be able to ask the video player "What am I looking at here, describe the equations" and it will OCR the frames, analyze them and explain them to me.

It's only a matter of time before "browsing" means navigating HTTP sites via LLM prompts. although, I think it is critical that LLM input should NOT be restricted to verbal cues. Not everyone is an extrovert that longs to hear the sound of their own voices. A lot of human communication is non-verbal.

Once we get over the privacy implications (and I do believe this can only be done by worldwide legislative efforts), I can imagine looking at a "website" or video, and my expressions, mannerisms and gestures will be considered prompts.

At least that is what I imagine the tech would evolve into in 5+ years.

  • groby_b 3 days ago

    Now? OK, you need to screencap and upload to LLM, but that's well established tech by now. (Where by "well established", I mean at least 9 months old ;)

    Same goes for "navigating HTTP sites via LLM prompts". Most LLMs have web search integration, and the "Deep Research" variants do more complex navigation.

    Video chat is there partially, as well. It doesn't really pay much attention to gestures & expressions, but I'd put the "earliest possible" threshold for that a good chunk closer than 5 years.

    • notepad0x90 3 days ago

      Yeah, all these things are possible today, but getting them well polished and integrated is another story. Imagine all this being supported by "HTML6" lol. When apple gets around to making this part of safari, then we know it's ready.

      • groby_b 3 days ago

        That's a great upper-bound estimator ;)

        But kidding aside - I'm not sure people want this being supported by web standards. We could be a huge step closer to that future had we decided to actually take RDF/Dublin Core/Microdata seriously. (LLMs perform a lot better with well-annotated data)

        The unanimous verdict across web publishers was "looks like a lot of work, let's not". That is, ultimately, why we need to jump through all the OCR hoops. Not only did the world not annotate the data, it then proceeded to remove as many traces of machine readability as possible.

        So, the likely gating factor is probably not Apple & Safari & "HTML6" (shudder!)

        If I venture my best bet what's preventing polished integration: It's really hard to do via foundational models only, and the number of people who want to have deep & well-informed conversations via a polished app enough that they're willing to pay for an app that does that is low enough that it's not the hot VC space. (Yet?)

        Crystal ball: Some OSS project will probably get within spitting distance of something really useful, but also probably flub the UX. Somebody else will take up these ideas while it's hot and polish it in a startup. So, 18-36 months for an integrated experience from here?

  • devmor 3 days ago

    Good lord, I dearly hope not. That sounds like a coddled hellscape world, something you'd see made fun of in Disney's Wall-E.

    • notepad0x90 3 days ago

      hence my comment about privacy and need for legislation :)

      It isn't the tech that's the problem but the people that will abuse it.

      • devmor 3 days ago

        While those are concerns, my point was that having everything on the internet navigated to, digested and explained to me sounds unpleasant and overall a drain on my ability to think and reason for myself.

        It is specifically how you describe using the tech that provokes a feeling of revulsion to me.

        • notepad0x90 3 days ago

          Then I think you misunderstand. The ML system would know when you want things digested to you or not. Right now companies are assuming this and forcing LLM interaction. But when properly done, the system would know based on your behavior or explicit prompts what you want and provide the service. If you're staring at a paragraph intently and confused, it might start highlighting common phrases or parts of the text/picture that might be hard to grasp and based on your reaction to that, it might start describing things via audio,tool tips,side pane,etc.. In other words, if you don't like how and when you're interacting with the LLM ecosystem, then that is an immature and failing ecosystem, in my vision this would be a largely solved problems, like how we interact with keyboards,mouse and touchscreens today.

          • devmor 3 days ago

            No, I fully understand.

            I am saying that this type of system, that deprives the user of problem solving, is itself a problem. A detriment to the very essence of human intelligence.

            • notepad0x90 2 days ago

              I just look at it as allowing the user to focus on problems that aren't already easily solved. Like using a calculator instead of calculating manually on paper.

              • devmor 2 days ago

                But the scenario you described is one in which you need an equation explained to you. That is exactly the kind of scenario where it's important to do the calculation yourself to understand it.

                If you are expecting problems to be solved for you, you are not learning, you're just consuming content.

  • abrichr 3 days ago

    > I wondered how soon will I be able to ask the video player "What am I looking at here, describe the equations" and it will OCR the frames, analyze them and explain them to me.

    Seems like https://aiscreenshot.app might fit the bill.

groby_b 3 days ago

Perusing the web site, it's depressing how much behind Mistral is on basic "how can I make this a compelling hook for customers" for the page.

The notebook link? An ACL'd doc

The examples don't even include a small text-to-markdown sample.

The before/after slider is cute, but useless - SxS is a much better way to compare.

Trying it in "Le Chat" requires a login.

It's like an example of "how can we implement maximum loss across our entire funnel". (I have no doubt the underlying tech does well, but... damn, why do you make it so hard to actually see it, Mistral?)

If anybody tried it and has shareable examples - can you post a link? Also, anybody tried it with handwriting yet?

michaelbuckbee 3 days ago

I'd mentioned this on HN last month, but I took a picture of a grocery list and then pasted it into ChatGPT to have it written out and it worked flawlessly...until I discovered that I'd messed up the picture when I took it at an angle and had accidentally cut off the first character or two of the bottom half of the list.

ChatGPT just inferred that I wanted the actual full names of the items (aka "flour" instead of "our").

Depending on how you feel about it, this is either an absolute failure of OCR or wildly useful and much better.

sunami-ai 3 days ago

Making Transformers the same cost as CNN's (which are used in character-level ocr, as opposed to image-patch-level) is a good thing. The problem with CNN based character-level OCR is not the recognition models but the detection models. In a former life, I found a way to increase detection accuracy, and, therefore, overall OCR accuracy, and used that as an enhancement on top of Amazon and Google OCR. It worked really well. But the transformer approach is more powerful and if it can be done for $1 per 1000 pages, that is a game changer, IMO, at least of incumbents offering traditional character-level OCR.

  • menaerus 3 days ago

    It certainly isn't the same cost if expressed as a non-subsidized $$$ one needs for the Transformers compute aka infra.

    CNNs trained specifically for OCR can run in real time on as small compute as a mobile device is.

    • anon373839 3 days ago

      A bit of a tangent, but aren’t CNNs still dominating over ViTs among computer vision competition winners?

      • menaerus 2 days ago

        I haven't watched that space very closely but IMO ViTs have a great potential to extract from since in comparison to CNNs they allow the model to learn and understand complex relations in the data. Where this matters, I expect it to matter a lot. OCR I think is not the greatest such example - while it matters to understand the surrounding context, I think it's not that critical for performance.

s4i 3 days ago

I wonder how good it would be to convert sheet music to MusicXML. All the current tools more or less suck with this task, or maybe I’m just ignorant and don’t know what lego bricks to put together.

TriangleEdge 3 days ago

One of my hobby projects while in University was to do OCR on book scans. Doing character recognition was solved, but finding the relationship between characters was very difficult. I tried "primitive" neural nets, but edge cases would often break what I built. Super cool to me to see such an order of magnitude in improvement here.

Does it do hand written notes and annotations? What about meta information like highlighting? I am also curious if LLMs will get better because more access to information if it can be effectively extracted from PDFs.

  • jcuenod 3 days ago

    * Character recognition on monolingual text in a narrow domain is solved

z2 3 days ago

Is there a reliable handwriting OCR benchmark out there (updated, not a blog post)? Despite the gains claimed for printed text, I found (anecdotally) that trying to use Mistral OCR on my messy cursive handwriting to be much less accurate than GPT-4o, in the ballpark of 30% wrong vs closer to 5% wrong for GPT-4o.

Edit: answered in another post: https://huggingface.co/spaces/echo840/ocrbench-leaderboard

InvidFlower 2 days ago

While it is nice to have more options, it still definitely isn't at a human level yet for hard to read text. Still haven't seen anything that can deal with something like this very well: https://i.imgur.com/n2sBFdJ.jpeg

If I remember right, Gemini actually was the closest as far as accuracy of the parts where it "behaved", but it'd start to go off the rails and reword things at the end of larger paragraphs. Maybe if the image was broken up into smaller chunks. In comparison, Mistral for the most part (besides on one particular line for some reason) sticks to the same number of words, but gets a lot wrong on the specifics.

qwertox 3 days ago

We developers seem to really dislike PDFs, to a degree that we'll build LLMs and have them translate it into Markdown.

Jokes aside, PDFs really serve a good purpose, but getting data out of them is usually really hard. They should have something like an embedded Markdown version with a JSON structure describing the layout, so that machines can easily digest the data they contain.

  • siva7 3 days ago

    > We developers seem to really dislike PDFs, to a degree that we'll build LLMs and have them translate it into Markdown.

    Why Jokes aside? Markdown/html is better suited for the web than pdf

egorfine 2 days ago

I had a need to scan serial numbers from Apple's product boxes out of pictures taken by a random person on their phone.

All OCR tools that I have tried have failed. Granted, I would get much better results if I used OpenCV to detect the label, rotate/correct it, normalize contrast, etc.

But... I have tried the then new vision model from OpenAI and it did the trick so well it's wasn't feasible to consider anything else at that point.

I have checked all S/N afterwards for being correct via third-party API - and all of theme were. Sure, sometimes I had to check versions with 0/o and i/l/1 substitutions but I believe these kind of mistakes are non-issues.

climb_stealth 3 days ago

Does this support Japanese? They list a table of language comparisons againat other approaches but I can't tell if it is exhaustive.

I'm hoping that something like this will be able to handle 3000-page Japanese car workshop manuals. Because traditional OCR really struggles with it. It has tables, graphics, text in graphics, the whole shebang.

protonbob 3 days ago

Wow this basically "solves" DRM for books as well as opening up the door for digitizing old texts more accurately.

bsnnkv 3 days ago

Someone working there has good taste to include a Nizar Qabbani poem.

hdjrudni 2 days ago

Still terrible at handwriting.

I signed up for the API, cobbled together from their tutorial (https://docs.mistral.ai/capabilities/document/) -- why can't they give the full script instead of little bits?

Tried uploading a tiff, they rejected it. Tried upload JPG, they rejected it (even though they supposed support images?). Tried resaving as PDF. It took that, but the output was just bad. Then tried ChatGPT on the original .tiff (not using API), and it got it perfectly. Honestly I could barely make out the handwriting with my eyes but now that I see ChatGPT's version I think it's right.

  • InvidFlower 2 days ago

    It is confusing, but they have diff calls for pdfs vs images. In their example google colab: https://colab.research.google.com/drive/11NdqWVwC_TtJyKT6cmu...

    The first couple of sections are for pdfs and you need to skip all that (search for "And Image files...") to find the image extraction portion. Basically it needs ImageURLChunk instead of DocumentURLChunk.

andoando 3 days ago

Bit unrelated but is there anything that can help with really low resolution text? My neighbor got hit and run the other day for example, and I've been trying every tool I can to make out some of the letters/numbers on the plate

https://ibb.co/mr8QSYnj

  • zinglersen 3 days ago

    Finding the right subreddit and asking there is probably a better approach if you want to maximize the chances of getting the plate 'decrypted'.

  • rvnx 3 days ago

    If it’s a video, sharing a few frames can help as well

  • dewey 3 days ago

    To even get started on this you'd also need to share some contextual information like continent, country etc. I'd say.

    • andoando 3 days ago

      Its in CA, looks like paper plates which follow a specific format and the last two seem to be the numbers '64'. Police should be able to search for temp tag with partial match and match the make/model. Was curious to see if any software could help though

  • busymom0 3 days ago

    There are photo enhancers online. But your picture is way too pixelated to get any useful info from it.

    • tjoff 3 days ago

      If you know the font in advance (which you often do in these cases) you can do insane reconstructions. Also keep in mind that it doesn't have to be a perfect match, with the help of the color and other facts (such as likely location) about the car you can narrow it down significantly.

    • zellyn 3 days ago

      Maybe if you had multiple frames, and used something very clever?

  • flutas 3 days ago

    Looks like a paper temp tag. Other than that, I'm not sure much can be had from it.

jacooper 3 days ago

Pretty cool, would love to use this with paperless, but I just can't bring myself to send a photo of all my documents to a third party, especially legal and sensitive documents, which is what I use Paperless for.

Because of that I'm stuck with crappy vision on Ollama (Thanks to AMDs crappy ROCm support for Vllm)

hubraumhugo 3 days ago

It will be interesting to see how all the companies in the document processing space adapt as OCR becomes a commodity.

The best products will be defined by everything "non-AI", like UX, performance and reliability at scale, and human-in-the loop feedback for domain experts.

  • trollied 3 days ago

    They will offer integrations into enterprise systems, just like they do today.

    Lots of big companies don't like change. The existing document processing companies will just silently start using this sort of service to up their game, and keep their existing relationships.

  • hyuuu 3 days ago

    I 100% agree with this, I think you can even extend this to any AI, in the end, IMO, as the llm is more commoditized, the surface of which the value is delivered will matter more

mjnews 2 days ago

> Mistral OCR has shown impressive performance, but OCR remains a challenging problem, especially with the risk of hallucinations and missing text in LLM-based approaches. For those interested in exploring its capabilities further, the official site provides more details: [Mistral OCR](https://www.mistralocr.org). It would be great to see more benchmarks comparing different OCR solutions in real-world scenarios.

sixhobbits 3 days ago

Nice demos but I wonder how well it does on longer files. I've been experimenting with passing some fairly neat PDFs to various LLMs for data extraction. They're created from Excel exports and some of the data is cut off or badly laid out, but it's all digitally extractable.

The challenge isn't so much the OCR part, but just the length. After one page the LLMs get "lazy" and just skip bits or stop entirely.

And page by page isn't trivial as header rows are repeated or missing etc.

So far my experience has definitely been that the last 2% of the content still takes the most time to accurately extract for large messy documents, and LLMs still don't seem to have a one-shot solve for that. Maybe this is it?

  • hack_ml 3 days ago

    You will have to send one page at a time, most of this work has to be done via RAG. Adding a large context (like a whole PDF), still does not work that well in my experience.

dotnetkow 3 days ago

Congrats to the Mistral team for launching! A general-purpose OCR model is useful, of course. However, more purpose-built solutions are a must to convert business documents reliably. AI models pre-trained on specific document types perform better and are more accurate. Coming soon from the ABBYY team, we're shipping a new OCR API designed to be consistent, reliable, and hallucination-free. Check it out if you're looking for best-in-class DX: https://digital.abbyy.com/code-extract-automate-your-new-mus...

strangescript 2 days ago

I think its interesting they left out Gemini 2.0 Pro in the benchmarks which I find to be markedly better than flash if you don't mind the spend.

lingjiekong 2 days ago

Curious that have people find more details regarding what is the architecture of this "mistral-ocr-latest". I have two question that

1. I was initially thinking this is VLM parsing model until I saw it can extract images. Then, I assume it is a pipeline of an image extraction and a VLM model while their result is combined to give the final result.

2. In this case, benchmark the pipeline result vs a end to end VLM such as gemini 2.0 flash might not be apple to apple comparison.

pqdbr 3 days ago

I tried with both PDFs and PNGs in Le Chat and the results were the worst I've ever seen when compared to any other model (Claude, ChatGPT, Gemini).

So bad that I think I need to enable the OCR function somehow, but couldn't find it.

  • troyvit 3 days ago

    It worked perfectly for me with a simple 2 page PDF that contained no graphics or formatting beyond headers and list items. Since it was so small I had the time to proof-read it and there were no errors. It added some formatting, such as bolding headers in list items and putting tics around file and function names. I won't complain.

  • computergert 3 days ago

    I'm experiencing the same. Maybe the sentence "Mistral OCR capabilities are free to try on le Chat." was a hallucination.

bob1029 3 days ago

> It takes images and PDFs as input

If you are working with PDF, I would suggest a hybrid process.

It is feasible to extract information with 100% accuracy from PDFs that were generated using the mappable acrofields approach. In many domains, you have a fixed set of forms you need to process and this can be leveraged to build a custom tool for extracting the data.

Only if the PDFs are unknown or were created by way of a cellphone camera, multifunction office device, etc should you need to reach for OCR.

The moment you need to use this kind of technology you are in a completely different regime of what the business will (should) tolerate.

  • themanmaran 3 days ago

    > Only if the PDFs are unknown or were created by way of a cellphone camera, multifunction office device, etc should you need to reach for OCR.

    It's always safer to OCR on every file. Sometimes you'll have a "clean" pdf that has a screenshot of an Excel table. Or a scanned image that has already been OCR'd by a lower quality tool (like the built in Adobe OCR). And if you rely on this you're going to get pretty unpredictable results.

    It's way easier (and more standardized) to run OCR on every file, rather than trying to guess at the contents based on the metadata.

    • bob1029 3 days ago

      It's not guessing if the form is known and you can read the information directly.

      This is a common scenario at many banks. You can expect nearly perfect metadata for anything pushed into their document storage system within the last decade.

      • themanmaran 3 days ago

        Oh yea if the form is known and standardized everything is a lot easier.

        But we work with banks on our side, and one of the most common scenarios is customers uploading financials/bills/statements from 1000's of different providers. In which case it's impossible to know every format in advance.

kapitalx 3 days ago

Co-founder of doctly.ai here (OCR tool)

I love mistral and what they do. I got really excited about this, but a little disappointed after my first few tests.

I tried a complex table that we use as a first test of any new model, and Mistral OCR decided the entire table should just be extracted as an 'image' and returned this markdown:

``` ![img-0.jpeg](img-0.jpeg) ```

I'll keep testing, but so far, very disappointing :(

This document I try is the entire reason we created Doctly to begin with. We needed an OCR tool for regulatory documents we use and nothing could really give us the right data.

Doctly uses a judge, OCRs a document against multiple LLMs and decides which one to pick. It will continue to run the page until the judge scores above a certain score.

I would have loved to add this into the judge list, but might have to skip it.

  • bambax 3 days ago

    Where did you test it? At the end of the post they say:

    > Mistral OCR capabilities are free to try on le Chat

    but when asked, Le Chat responds:

    > can you do ocr?

    > I don't have the capability to perform Optical Character Recognition (OCR) directly. However, if you have an image with text that you need to extract, you can describe the text or provide details, and I can help you with any information or analysis related to that text. If you need OCR functionality, you might need to use a specialized tool or service designed for that purpose.

    Edit: Tried anyway by attaching an image; it said it could do OCR and then output... completely random text that had absolutely nothing to do with the text in the image!... Concerning.

    Tried again with a better definition image, output only the first twenty words or so of the page.

    Did you try using the API?

    • kapitalx 3 days ago

      Yes I used the API. They have examples here:

      https://docs.mistral.ai/capabilities/document/

      I used base64 encoding of the image of the pdf page. The output was an object that has the markdown, and coordinates for the images:

      [OCRPageObject(index=0, markdown='![img-0.jpeg](img-0.jpeg)', images=[OCRImageObject(id='img-0.jpeg', top_left_x=140, top_left_y=65, bottom_right_x=2136, bottom_right_y=1635, image_base64=None)], dimensions=OCRPageDimensions(dpi=200, height=1778, width=2300))] model='mistral-ocr-2503-completion' usage_info=OCRUsageInfo(pages_processed=1, doc_size_bytes=634209)

      • sadcrab 2 days ago

        Any luck with this? I'm trying to process photos of paperwork (.pdf, .png) and got the same results as you.

        Feels like something is missing in the docs, or the API itself.

        https://imgur.com/a/1J9bkml

  • fnordpiglet 3 days ago

    Interestingly I’m currently going through and scanning the hundreds of journal papers my grandfather authored in medicine and thinking through what to do about graphs. I was expecting to do some form of multiphase agent based generation of LaTeX or SVG rather than a verbal summary of the graphs. At least in his generation of authorship his papers clearly explained the graphs already. I was pretty excited to see your post naturally but when I looked at the examples what I saw was, effectively, a more verbose form of

    ``` ![img-0.jpeg](img-0.jpeg) ```

    I’m assuming this is partially because your use case is targeting RAG under various assumptions bur also partially because multimodal models aren’t near what I would need to be successful with?

    • kapitalx 3 days ago

      We need to update the examples on the front page. Currently for things that are considered charts/graphs/figures we convert to a description. For things like logos or images we do an image tag. You can also choose to exclude them.

      The difference with this is that it took the entire page as an image tag (it's just a table of text in my document). rather than being more selective.

      I do like that they give you coordinates for the images though, we need to do something like that.

      Give the actual tool a try. Would love to get your feedback for that use case. It gives you 100 free credits initially but if you email me (ali@doctly.ai), I can give you an extra 500 (goes for anyone else here also)

  • niwtsol 3 days ago

    If you have a judge system, and Mistral performs well on other tests, wouldn't you want to include it so if it scores the highest by your judges ranking it would select the most accurate result? Or are you saying that mistral's image markdown would score higher on your judge score?

    • kapitalx 3 days ago

      We'll definitely be doing more tests, but the results I got on the complex tests would result in a lower score and might not be worth the extra cost of the judgement itself.

      In our current setup Gemini wins most often. We enter multiple generations from each model into the 'tournament', sometimes one generation from gemini could be at the top while another in the bottom, for the same tournament.

  • Grosvenor 3 days ago

    Does doctly do handwritten forms like dates?

    I have a lot of "This document filed and registered in the county of ______ on ______ of _____ 2023" sort of thing.

    • kapitalx 3 days ago

      We've been getting great results with those aswell. But ofcourse there is always some chance of not getting it perfect, specially with different handwritings.

      Give it a try, no credit cards needed to try it. If you email me (ali@doctly.ai) i can give you extra free credits for testing.

      • Grosvenor 3 days ago

        Just tried it. Got all the dates correct and even extracted signatures really well.

        Now to figure out how many millions of pages I have.

  • scottydelta 3 days ago

    How do you stay competitive with $2/100 pages pricing as compared to mistral and others offering 1000 pages for $1 approx?

    • kapitalx 9 hours ago

      Customers are willing to pay for accuracy compared to existing solutions out there. We started out in need of an accurate solution for a RAG product we were building, but none of the solutions we tried were providing the accuracy we needed.

  • infecto 3 days ago

    Why pay more for doctly than an AWS Textract?

    • kapitalx 3 days ago

      Great question. The language models are definitely beating the old tools. Take a look at Gemini for example.

      Doctly runs a tournament style judge. It will run multiple generations across LLMs and pick the best one. Outperforming single generation and single model.

  • the_mitsuhiko 3 days ago

    Would love to see the test file.

    • Starlord2048 3 days ago

      would be glad to see benchmarking results

      • kapitalx 3 days ago

        This is a good idea. We should publish a benchmark results/comparison.

jervant 3 days ago

I wonder how it compares to USPS workers at deciphering illegible handwriting.

Oras 3 days ago

I feel this is created for RAG. I tried a document [0] that I tested with OCR; it got all the table values correctly, but the page's footer was missing.

Headers and footers are a real pain with RAG applications, as they are not required, and most OCR or PDF parsers will return them, and there is extract work to do to remove them.

[0] https://github.com/orasik/parsevision/blob/main/example/Mult...

soyyo 2 days ago

I understand that is more juicy to get information from graphs, figures and so on, as every domain uses those, but i really hope to eventually see these models to be able to workout music notation, i have tried the best known apps and all of them fail to capture important details such as guitar performace symbols for bends or legato

pawelduda 3 days ago

It outperforms the competition significantly AND can extract embedded images from the text. I really like LLMs for OCR more and more. Gemini was already pretty good at it

coolspot 3 days ago

This is $1 per 1000 pages. For comparison, Azure Document Intelligence is $1.5/1000 pages for general OCR and $30/1000 pages for “custom extraction”.

  • 0cf8612b2e1e 3 days ago

    Given the wide variety of pricing on all of these providers, I keep wondering how the economics work. Do they have fantastic margin on some of these products or is it a matter of subsidizing the costs, hoping to capture the market? Last I heard, OpenAI is still losing money.

thomasahle 2 days ago

I'm surprised they didn't benchmark it against Pixtral.

They test it against a bunch of different Multimodal LLMs, so why not their own?

I don't really see the purpose of the OCR form factor, when you have multimodal LLMs. Unless it's significantly cheaper.

peterburkimsher 3 days ago

Does it work for video subtitles? And in Chinese? I’m looking to transcribe subtitles of live music recordings from ANHOP and KHOP.

sureglymop 3 days ago

Looks good but in the first hover/slider demo one can see how it could lead to confusion when handling side by side content.

Table 1 is referred to in section `2 Architectural details` but before `2.1 Multimodal Decoder`. In the generated markdown though it is below the latter section, as if it was in/part of that section.

Of course I am nitpicking here but just the first thing I noticed.

  • 0cf8612b2e1e 3 days ago

    Does anything handle dual columns well? Despite being the academic standard, it seemingly throws off every generic tool.

th0ma5 3 days ago

A great question for people wanting to use OCR in business is... Which digits in monetary amounts can you tolerate being incorrect?

rvz 3 days ago

> "Fastest in its category"

Not one mention of the company that they have partnered with and that is Cerebras AI and that is the reason they have fast inference [0]

Literally no-one here is talking about them and they are about to IPO.

[0] https://cerebras.ai/blog/mistral-le-chat

low_tech_punk 3 days ago

This might be a contrarian take: the improvement against gpt-4o and gemini-1.5 flash, both of which are general purpose multi-modal models, seem to be underwhelming.

I'm sensing another bitter lesson coming, where domain optimized AI will hold a short term advantage but will be outdated quickly as the frontier model advances.

roboben 3 days ago

Le chat doesn’t seem to know about this change despite the blog post stating it. Can anyone explain how to use it in Le Chat?

submeta 3 days ago

Is this able to convert pdf flowcharts into yaml or json representations of them? I have been experimenting with Claude 3.5. It has been very good at readig / understanding/ converting into representations of flow charts.

So I am wondering if this is more capable. Will try definitely, but maybe someone can chime in.

gatienboquet 3 days ago

I feel like i can't create an agent with their OCR model yet ? Is it something planned or it's only API?

jcuenod 3 days ago

Just tested with a multilingual (bidi) English/Hebrew document.

The Hebrew output had no correspondence to the text whatsoever (in context, there was an English translation, and the Hebrew produced was a back-translation of that).

Their benchmark results are impressive, don't get me wrong. But I'm a little disappointed. I often read multilingual document scans in the humanities. Multilingual (and esp. bidi) OCR is challenging, and I'm always looking for a better solution for a side-project I'm working on (fixpdfs.com).

Also, I thought OCR implied that you could get bounding boxes for text (and reconstruct a text layer on a scan, for example). Am I wrong, or is this term just overloaded, now?

  • nicodjimenez 3 days ago

    You can get bounding boxes from our pdf api at Mathpix.com

    Disclaimer, I’m the founder

    • kergonath 3 days ago

      Mathpix is ace. That’s the best results I got so far for scientific papers and reports. It understands the layout of complex documents very well, it’s quite impressive. Equations are perfect, figures extraction works well.

      There are a few annoying issues, but overall I am very happy with it.

      • nicodjimenez 3 days ago

        Thanks for the kind words. What are some of the annoying issues?

        • kergonath a day ago

          I had a billing issue at the beginning. It was resolved very nicely but I try to be careful and I monitor the bill a bit more than I would like.

          Actually my main remaining technical issue is conversion to standard Markdown for use in a data processing pipeline that has issues with the Mathpix dialect. Ideally I’d do it on a computer that is airgaped for security reasons. But I haven’t found a very good way of doing it because the Python library wanted to check my API key.

          A problem I have and that is not really Mathpix’s fault is that I don’t really know how to store the figures pictures to keep them with the text in a convenient way. I haven’t found a very satisfying strategy.

          Anyway, keep up the good work!

pilooch 3 days ago

But what's the need exactly for OCR when you have multimodal LLMs that can read the same info and directly answer any questions about it ?

For a VLLM, my understanding is that OCR corresponds to a sub-field of questions, of the type 'read exactly what's written in this document'.

  • simonw 3 days ago

    The biggest risk of vision LLMs for OCR is that they might accidentally follow instructions is the text that they are meant to be processing.

    (I asked Mistral if their OCR system was vulnerable to this and they said "should be robust, but curious to see if you find any fun examples" - https://twitter.com/simonw/status/1897713755741368434 and https://twitter.com/sophiamyang/status/1897719199595720722 )

    • pilooch 3 days ago

      Fun, but LLMs would follow them post OCR anyways ;)

      I see OCR much like phonemes in speech, once you have end to end systems, they become latent constructs from the past.

      And that is actually good, more code going into models instead.

  • daemonologist 3 days ago

    It's useful to have the plain text down the line for operations not involving a language model (e.g. search). Also if you have a bunch of prompts you want to run it's potentially cheaper, although perhaps less accurate, to run the OCR once and save yourself some tokens or even use a smaller model for subsequent prompts.

  • ks2048 3 days ago

    Tons of uses: Storage (text instead of images), search (user typing in a text box and you want instant retrieval from a dataset), etc. And costs: run on images once - then the rest of your queries will only need to run on text.

dwedge 2 days ago

Benchmarks look good. I tried this with a PDF that already has accurate PDF embedded just with new lines making pdftotext fail, and it was accurate for the text it found, but missed entire pages

alberth 3 days ago

Curious to see how this performance against more real world usage of someone taking a photo of text (which the text then becomes slightly blurred) and performing OCR on it.

I can't exactly tell if the "Mistral 7B" image is an example of this exact scenario.

monkeydust 2 days ago

Spent time working on OCR problem many years ago for a mobile app. We found at the time that the preprocessing was so critical to the outcome (quality of image, angle, colour/greyscale)

101008 3 days ago

Is this free in LeChat? I uploaded a handwritten text and it stopped after the 4th word.

riffic 3 days ago

It'd be great if this could be tested against genealogical documents written in cursive like oh most of the documents on microfilm stored by the LDS on familysearch, or eastern european archival projects etc.

Gnan 3 days ago

Is there an ocr with this kind of accuracy, but can run in a mobile device ? Looking for an ocr that can detect texts with high accuracy in realtime, so option of using cloud ocr is not viable.

lokl 3 days ago

Tried with a few historical handwritten German documents, accuracy was abysmal.

  • lysace 3 days ago

    Semi-OT (similar language): The national archives in Sweden and Finland published a model for OCR:ing handwritten Swedish text from the 1600s to the 1800s with what to me seems like a very level of accuracy given the source material. (4% character error rate)

    https://readcoop.eu/model/the-swedish-lion-i/

    https://www.transkribus.org/success-story/creating-the-swedi...

    https://huggingface.co/Riksarkivet

    They have also published a fairly large volume of OCR:ed texts (IIRC birth/death notices from church records) using this model online. As a beginner genealogist it's been fun to follow.

  • Thaxll 3 days ago

    HTR ( Handwritten Text Recognition ) is a completely different space than OCR. What were you expecting exactly?

    • riquito 3 days ago

      It fits the "use cases" mentioned in the article

      > Preserving historical and cultural heritage: Organizations and nonprofits that are custodians of heritage have been using Mistral OCR to digitize historical documents and artifacts, ensuring their preservation and making them accessible to a broader audience.

      • Thaxll 3 days ago

        There is a difference between historical document and "my doctor prescription".

        Someone coming here and saying it does not work with my old german hanwriting doesn't say much.

        • riquito 3 days ago

          You're making a strawman, the parent specifically mentioned "historical handwritten documents"

  • butovchenkoy 2 days ago

    For this task, general models will always perform poorly. My company trains custom gen ai models for document understanding. We recently trained a VLM for the German government to recognize documents written in old German handwriting, and it performed with exceptionally high accuracy.

  • rvnx 3 days ago

    Probably they are overfitting the benchmarks, since other users also complain of the low accuracy

  • thadt 3 days ago

    Also working with historical handwritten German documents. So far Gemini seems to be the least wrong of the ones I've tried - any recommendations?

    • butovchenkoy 2 days ago

      my recommendation is to train a custom model

  • anothermathbozo 3 days ago

    Optical Character Recognition (OCR) and Handwritten Text Recognition (HTR) are different tasks

joeevans1000 3 days ago

I've found that the stunning OCR results so far were because the models were trained on the example file category. Is that the case here? Or can this recognize various documents?

  • noosphr 3 days ago

    After four years of "This VLM will solve OCR forever!" posts I've firmly put VLMs in the "useless until 100t parameters" category.

    Some day when there is enough internal state and training data that they can recognize tables, images and text we will get a gpt3 like moment that will make regular OCR obsolete.

    But that day is very far off and everyone who I've talked with and consulted over using VLMs in their pipeline is better served doing something else specific to their use case.

    An example I've used before: https://imgur.com/a/w2XKObh

    The reply:

    ---

    To extract the text contained within each box while ignoring the text in sub-boxes, we can follow these steps:

    1. Identify the outermost box and extract its text. 2. Move to the next outermost box and extract its text, ignoring any text within its sub-boxes. 3. Continue this process for all boxes.

    Let's apply this to the image provided:

    1. The outermost box contains the text: "This should be second". 2. The next outermost box (ignoring the sub-box within it) contains the text: "First". 3. The next box contains the text: "And also this". 4. The final box contains the text: "The quick brown fox".

    So, the extracted text from each box, ignoring sub-boxes, is:

    1. "This should be second" 2. "First" 3. "And also this" 4. "The quick brown fox"

    ---

    As you can plainly see it is _wildly_ wrong and gives you no way to try and recover from those errors.

atemerev 3 days ago

So, the only thing that stopped AI from learning from all our science and taking over the world was the difficulty of converting PDFs of academic papers to more computer readable formats.

Not anymore.

kinnth 2 days ago

This looks like a massive win if you were the NHS and had to scan and process old case notes.

Same is true if you were a solicitors/lawyers.

kccqzy 3 days ago

I have an actually hard OCR exercise for an AI model: I take this image of Chinese text on one of the memorial stones on the Washington Monument https://www.nps.gov/articles/american-mission-ningpo-china-2... and ask the model to do OCR. Not a single model I've seen can OCR this correctly. Mistral is especially bad here: it gets stuck in an endless loop of nonsensical hallucinated text. Insofar as Mistral is design for "preserving historical and cultural heritage" it couldn't do that very well yet.

A good model can recognize that the text is written top to bottom and then right to left and perform OCR in that direction. Apple's Live Text can do that, though it makes plenty of mistakes otherwise. Mistral is far from that.

newfocogi 3 days ago

They say: "releasing the API mistral-ocr-latest at 1000 pages / $"

I had to reread that a few times. I assume this means 1000pg/$1 but I'm still not sure about it.

  • dgfl 3 days ago

    Great example of how information is sometimes compartmentalized arbitrarily in the brain: I imagine you have never been confused by sentences such as “I’m running at 10 km/h”.

    • mkl 3 days ago

      Dollar signs go before the number, not after it like units. It needs to be 1000 pages/$1 to make sense, whereas 10km and 10h and 10/h all make sense so 10km/h does. I imagine you would be confused by km/h 10 but not $10.

      • ekwav 2 days ago

        In the EU we put the € symbol after the number so it feels more natural to do that for $ as well.

  • svachalek 3 days ago

    Yeah you can read it as "pages per dollar" or as a unit "pages/$", it all comes out the same meaning.

  • amelius 3 days ago

    Hmm, can it read small print? ;)

  • bredren 3 days ago

    Ya, presumably it is missing the number `1.00`.

    • groby_b 3 days ago

      Not really. When you go 60 mph (or km/h) you don't specify the 1.00 for the hours either. pages/$ is the unit, 1000 is the value.

      • bredren 3 days ago

        But you do for virtually all other cloud pricing pages.

shmoogy 3 days ago

What's the general time for something like this to hit openrouter? I really hate having accounts everywhere when I'm trying to test new things.

srinathkrishna 3 days ago

Given the fact that multi-modal LLMs are getting so good at OCR these days, is it a shame that we can't do local OCR with high accuracy in the near-term?

deadbabe 3 days ago

LLM based OCR is a disaster, great potential for hallucinations and no estimate of confidence. Results might seem promising but you’ll always be wondering.

  • menaerus 3 days ago

    CNN-based OCR also have "hallucinations" and Transformers aren't that much different in that respect. This is a problem solved with domain specific post-processing.

  • leumon 3 days ago

    well already in 2013 ocr systems used in xerox scanners (turned on by default!) randomly altered numbers, so its not an issue only occuring in llms.

thiago_fm 3 days ago

For general use this will be good.

But I bet that simple ML will lead to better OCRs when you are doing anything specialized, such as, medical documents, invoices etc.

coolspot 3 days ago

This is $1 per 1000 pages.

For comparison, Azure Document Intelligence is $1.5/1000 pages for general OCR and $30/1000 pages for “custom extraction”.

applgo443 3 days ago

What's the simple explanation for why these VLM OCRs hallucinate but previous version of OCRs don't?

  • prats226 2 days ago

    Traditional OCR's usually have detection + recognition pipeline. So they will detect every word and then try to predict the text for every word. Errors obviously can happen in both parts, eg some words not detected which will get missed from output. Or word recognized incorrectly which is also common and more comparable to hallucination. However give its trained to work only on a small patch, accuracy is often higher. Comparing this to VLM's, they are looking at entire image/context and auto-regressively generating tokens/text which can also have lot of language bias, hence hallicinations.

jslezak 3 days ago

Has anyone tried it for handwriting?

So far Gemini is the only model I can get decent output from for a particular hard handwriting task

thegabriele 3 days ago

I'm using gemini to solve textual CAPTCHA with some good results (better than untrained OCR).

I will give this a shot

nyeah 3 days ago

It's not fair to call it a "Mistrial" just because it hallucinates a little bit.

  • wendyshu 3 days ago

    Who called it that?

    • nyeah 2 days ago

      Well ... nobody. I read it wrong at first glance.

dehrmann 3 days ago

Is this burying the lede? OCR is a solved problem, but structuring document data from scans isn't.

anovick 3 days ago

How does one use it to identify bounding rectangles of images/diagrams in the PDF?

OrvalWintermute 3 days ago

I'm happy to see this development after being underwhelmed with Chatgpt OCR!

beebaween 3 days ago

Wonder how it does with table data in pdfs / page-long tabular data?

cavisne 3 days ago

Its funny how Gemini consistently beats googles dedicated document API.

  • jjice 3 days ago

    I'm not surprised honestly - it's just the newer better things vs their older offering

t_sea 2 days ago

They really went for it with the hieroglyphs opening.

d_llon 3 days ago

It's disappointing to see that the benchmark results are so opaque. I hope we see reproducible results soon, and hopefully from Mistral themselves.

1. We don't know what the evaluation setup is. It's very possible that the ranking would be different with a bit of prompt engineering.

2. We don't know how large each dataset is (or even how the metrics are calculated/aggregated). The metrics are all reported as XY.ZW%, but it's very possible that the .ZW% -- or even Y.ZW% -- is just noise.[1]

3. We don't know how the datasets were mined or filtered. Mistral could have (even accidentally!) filtered out particularly data points that their model struggled with. (E.g., imagine good-meaning engineer testing a document with Mistral OCR first, finding it doesn't work, and deducing that it's probably bad data and removing it.)

[1] https://medium.com/towards-data-science/digit-significance-i...

jbverschoor 3 days ago

Ohhh. Gonna test it out with some 100+ year old scribbles :)

  • jbverschoor 2 days ago

    It did better than any other solution out there. However, I can only validate by the logic of the text. It's a recipe book.

WhitneyLand 3 days ago

1. There’s no simple page / sandbox to upload images and try it. Fine, I’ll code it up.

2. “Explore the Mistral AI APIs” (https://docs.mistral.ai) links to all apis except OCR.

3. The docs on the api params refer to document chunking and image chunking but no details on how their chunking works?

So much unnecessary friction smh.

  • cooperaustinj 3 days ago

    There is an OCR page on the link you provided. It includes a very, very simple curl command (like most of their docs).

    I think the friction here exists outside of Mistral's control.

    • WhitneyLand 3 days ago

      How is it out of their control to document what they mean by chunking in their parameters?

    • kergonath 3 days ago

      > There is an OCR page on the link you provided.

      I don’t see it either. There might be some caching issue.

noloz 3 days ago

Are there any open source projects with the same goal?

ritvikpandey21 3 days ago

as builders in this space, we decided to put it to the test on complex nested tables, pie charts, etc. to see if the same VLM hallucination issues persist, and to what degree. while results were promising, we found several critical failure nodes across two document domains.

check out our blog post here! https://www.runpulse.com/blog/beyond-the-hype-real-world-tes...

revskill 2 days ago

Nextjs error is still uncauht correctly.

linklater12 3 days ago

Document processing is where b2b SAAS is at.

jwr 3 days ago

Alas, I can't run it locally. So it still doesn't solve the problem of OCR for my PDF archive containing my private data...

zelcon 2 days ago

Release the weights or buy an ad

maCDzP 3 days ago

Oh - on premise solution - awesome!

rjurney 3 days ago

What about tables in PDFs?

Zopieux 3 days ago

Saving you a click: no, it cannot be self hosted (unless you have a few million dollars laying around)

ein0p 2 days ago

Could anyone suggest a tool which would take a bunch of PDFs (already OCR-d with Finereader), and replace the OCR overlay on all of them, maintaining the positions? I would like to have more accurate search over my document archive.

jhatemyjob 2 days ago

As far as open source OCRs go, Tesseract is still the best, right?

bugglebeetle 3 days ago

Congrats to Mistral for yet again releasing another closed source thing that costs more than running an open source equivalent:

https://github.com/DS4SD/docling

  • Squarex 3 days ago

    I am all for open source, but where do you see benchmarks that conclude that it's just equivalent?

    • bugglebeetle 3 days ago

      Where do you see open source benchmark results that confirm Mistral’s performance?

  • anonymousd3vil 3 days ago

    Back in my days Mistral used to torrent models.

joeevans1000 3 days ago

Can someone give me a tl&dr on how to start using this? Is this available if one signs up for a regular Mistral account?

kiratp 3 days ago

It's shocking how much our industry fails to see past its own nose.

Not a single example on that page is a Purchase Order, Invoice etc. Not a single example shown is relevant to industry at scale.

  • merb 3 days ago

    Mistral is Europe based where invoices are more or less sent digitally in like 95% of all the cases anyway. Some are even digital invoices, which will at some point in the eu be mandatory. For orders there are proposals for that, too. And basically invoice data extraction is a different beast.

    • codetrotter 3 days ago

      One use-case is digitising receipts from business related travels for expenses that employees paid for out of their own pocket and which they are submitting pictures to the business for reimbursement.

      Bus travels, meals including dinners and snacks, etc. for which the employee has receipts on paper.

      • merb 2 days ago

        Receipts are different. And they are harder to OCR. Thermo prints are most often aweful in quality. Most often you need to correct some stuff when dealing with them. I doubt that this tech changes that significantly.

    • revnode 3 days ago

      So an invoice attached to an email as a PDF is sent digitally ... those unfamiliar with PDF will think text and data extraction is trivial then, but this isn't true. You can have a fully digital, non-image PDF that is vector based and has what looks like text, but doesn't have a single piece of extractable text in it. It's all about how the PDF was generated. Tables can be formatted in a million ways, etc.

      Your best bet is to always convert it to an image and OCR it to extract structured data.

      • merb 2 days ago

        This is simply not true. Maybe it’s easier and you do not need 100% precision. But it is actually possible to extract text and layout of digital pdfs. Else it would be impossible to display it. Of course some people still add image fragments to a pdf, but that practice is basically dying. I did not see a single pdf the last year we‘re it was impossible to extract the layout.

    • wolfi1 3 days ago

      even in Europe this is still a thing, I know of systems which still are unable to read items having more than one line (costing s sh*tload of money)

    • kiratp 3 days ago

      This isn't even close to true.

      Source: We have large EU customers.

      • merb 2 days ago

        So your eu customer will send you the invoice via letters ? Wow. There are some companies that still deal with printed invoices, but they are most often smaller companies that deal with health related things.

    • napolux 3 days ago

      Can confirm, in Italy electronic invoicing is mandatory since 2019

  • simpaticoder 3 days ago

    Another good example would be contracts of any kind. Imagine photographing a contract (like a car loan) and on the spot getting an AI to read it, understand it, forecast scenarious, highlight red flags, and do some comparison shopping for you.

    • JBiserkov 3 days ago

      ... imagining ...

      ... hallucinating during read ...

      ... hallucinating during understand ...

      ... hallucinating during forecast ...

      ... highlighting a hallucination as red flag ...

      ... missing an actual red flag ...

      ... consuming water to cool myself...

      Phew, being an AI is hard!

      • simpaticoder 3 days ago

        Your points are well-taken, but I think that contracts are a small enough, and well represented enough in the corpus, to actually be pretty solid. This is especially true with good prompting and some sort of feedback loop.

  • sha16 3 days ago

    I wanted to apply OCR to my company's invoicing since they basically did purchasing for a bunch of other large companies, but the variability in the conversion was not tolerable. Even rounding something differently could catch an accountant's eye, let alone detecting a "8" as a "0" or worse.

  • guiomie 3 days ago

    Agreed. In general I've had such bad performance for complex table based invoice parsing, that every few months I try the latest models to see if its better. It does say "96.12" on top-tier benchmark under the Table category.

  • arpinum 3 days ago

    Businesses at scale use EDI to handle purchase orders and invoices, no OCR needed.

    • cdolan 3 days ago

      Thats simply not a factual statement.

      Scaled businesses do USE edi, but they still receive hundreds of thousands of PDF documents a month

      source: built a saas product that handles pdfs for a specific industry

  • dotnetkow 3 days ago

    Agreed, though in this case, they are going for general-purpose OCR. That's fine in some cases, but purpose-built models trained on receipts, invoices, tax documents, etc., definitely perform better. We've got a similar API solution coming out soon (https://digital.abbyy.com/code-extract-automate-your-new-mus...) that should work better for businesses automating their docs at scale.

  • mtillman 3 days ago

    We find CV models to be better (higher midpoint on an ROC curve) for the types of docs you mention.

  • mentalgear 3 days ago

    To be fair: Reading the blog post, the main objective seems to have been to enable information extraction with high confidence for the academic sector (e.g. unlocking all these paper pdfs), and not necessarily to be another receipt scanner.

    • kiratp 3 days ago

      It hilarious that the academic sector 1. publishes as PDF 2. spends all this energy on how to extract that info back from PDF 3. publishes that research as PDF as well.

      Receipt scanning is a multiple orders of magnitude more valuable business. Mistral at this point is looking for a commercial niche (like how Claude is aiming at software development)

bondolo 3 days ago

Such a shame that PDF doesn’t just, like, include the semantic structure of the document by default. It is brilliant that we standardized on an archival document format that doesn’t include direct access to the document text or structure as a core intrinsic default feature.

I say this with great anger as someone who works in accessibility and has had PDF as a thorn in my side for 30 years.

  • NeutralForest 3 days ago

    I agree with this so much. I've tried to sometimes push friends and family to use text formats (at least I sent them something like Markdown), which is very easy to render in the browser anyways. But often you have to fall back to PDF, which I dislike very much. There's so much content like books and papers that are in PDF as well. Why did we pick a binary blob as shareable format again?

    • meatmanek 3 days ago

      > Why did we pick a binary blob as shareable format again?

      PDF was created to solve the problem of being able to render a document the same way on different computers, and it mostly achieved that goal. Editable formats like .doc, .html, .rtf were unreliable -- different software would produce different results, and even if two computers have the exact same version of Microsoft Word, they might render differently because they have different fonts available. PDFs embed the fonts needed for the document, and specify exactly where each character goes, so they're fully self-contained.

      After Acrobat Reader became free with version 2 in 1994, everybody with a computer ended up downloading it after running across a PDF they needed to view. As it became more common for people to be able to view PDFs, it became more convenient to produce PDFs when you needed everybody to be able to view your document consistently. Eventually, the ability to produce PDFs became free (with e.g. Office 2007 or Mac OS X's ability to print to PDF), which cemented PDF's popularity.

      Notably, the original goals of PDF had nothing to do with being able to copy text out of them -- the goal was simply to produce a perfect reproduction of the document on screen/paper. That wasn't enough of an inconvenience to prevent PDF from becoming popular. (Some people saw the inability for people to easily copy text from them as a benefit -- basically a weak form of text DRM.)

      • NeutralForest 2 days ago

        Thanks for the explanation! I was vaguely aware of those issues but not in depth. It all makes sense of course and now PDF is so deeply entrenched it's very difficult to push other formats. It's interesting that the contention between content and layout is still such an issue. I don't know what the fix is, maybe just the web?

  • cess11 3 days ago

    PDF is pretty strictly modeled on printed documents and their mainstream typography at the time of invention of Postscript and so on.

    Printed documents do not have any structure beyond the paper and placement of ink on them.

  • lukasb 3 days ago

    Even assuming you could get people to do the work (probably the real issue here) could a single schema syntax capture the semantics of the universe of documents that exist as PDFs? PDFs succeeded because they could reproduce anything.

  • andai 3 days ago

    Tables? I regularly run into PDFs where even the body text is mangled!

ChrisArchitect 3 days ago

[flagged]

  • vessenes 3 days ago

    No comments there yet - this at the top of the home page, let’s use this one.

hyuuu 3 days ago

It's a weird timing because I just launched https://dochq.io - ai document extraction where you can define what you need to get out your documents in plain English, I legitimately thought that this was going to be such a niche product but hell, there has been a very rapid rise for AI-based OCR lately, an article/tweet even went viral 2 weeks ago I think? About using Gemini to do OCR, fun times.