Announcement

Collapse
No announcement yet.

Tesseract 5.0 OCR Engine Bringing Faster Performance With "Fast Floats"

Collapse
X
 
  • Filter
  • Time
  • Show
Clear All
new posts

  • Tesseract 5.0 OCR Engine Bringing Faster Performance With "Fast Floats"

    Phoronix: Tesseract 5.0 OCR Engine Bringing Faster Performance With "Fast Floats"

    Tesseract as the leading open-source optical character recognition (OCR) engine that employs neural networks for converting images/scans of text into actual recognized text is nearing its 5.0 release...

    Phoronix, Linux Hardware Reviews, Linux hardware benchmarks, Linux server benchmarks, Linux benchmarking, Desktop Linux, Linux performance, Open Source graphics, Linux How To, Ubuntu benchmarks, Ubuntu hardware, Phoronix Test Suite

  • #2
    Curious about AVX512 optimizations, it makes a dramatic difference in libdav1d with 10 bit content (as Rocket Lake kills Zen 3 here). If this will be true for Tesseract too, its usefulness probably should be re-evaluated.

    Comment


    • #3
      LOL, I'm finishing a week long OCR run now. It's not obvious if fast float is enabled in the beta binaries or how to enable it in the build but maybe it's worth looking into...
      Last edited by elatllat; 16 August 2021, 02:42 PM.

      Comment


      • #4
        Traditionally the Tesseract OCR engine has relied upon doubles
        WTF??

        Did these clowns actually write their own deep learning framework? If so, they should ditch it and use one of the others, for probably an overnight order-of-magnitude speedup. Not exaggerating.
        Last edited by coder; 16 August 2021, 02:54 PM.

        Comment


        • #5
          Originally posted by aufkrawall View Post
          Curious about AVX512 optimizations, it makes a dramatic difference in libdav1d with 10 bit content (as Rocket Lake kills Zen 3 here).
          Look at other CPU-based deep learning benchmarks. OpenVINO is a good one, because both its AVX2 and AVX-512 paths are well-optimized. Michael has some other deep learning benchmarks in PTS that are pretty rubbish, in that they don't represent realistic performance on CPU.

          As for DAV1D, that's another rubbish comparison, simply because the non- AVX-512 path for 10-bit wasn't comparably optimized. So, it's not an apples-to-apples comparison.

          Comment


          • #6
            More optimization for non-AVX512 path in libdav1d is only a matter of time then? It's been a while since they started using it.

            Comment


            • #7
              So why were double instead of float used in the first place? Out of a habit?

              Comment


              • #8
                Originally posted by coder View Post
                WTF??

                Did these clowns actually write their own deep learning framework? If so, they should ditch it and use one of the others, for probably an overnight order-of-magnitude speedup. Not exaggerating.
                I don't know whether theirs is any good or not, but you should know that Tesseract was started in the 1980s, making it waaay older than any of the other things you're thinking of.

                Comment


                • #9
                  Tesseract always give me funny result when i OCR some japanese document

                  Comment


                  • #10
                    Originally posted by aufkrawall View Post
                    More optimization for non-AVX512 path in libdav1d is only a matter of time then? It's been a while since they started using it.
                    Someone probably has to fund the work. I forget who, but someone paid them to do the AVX-512 optimization of the 10-bit path. I think it was like Netflix or someone like that -- not Intel.

                    Comment

                    Working...
                    X