Hi,

I have some basic questions before starting a project of OCR
recognition for the iPhone.

I have seen the steps to cross-compile tesseract for iOS but have some
questions on tesseract roadmap itself:
1/ should I start on tesseract 2.4 or 3.0? From my understanding 3.0
is not yet stable but has a major refactoring ongoing plus several
features (including document layout analysis)? The current 3.0
"release" is quite far from the head of the trunk, which do not seem
to compile on iOS, so I am wondering if there is any new release
(3.01?) planned soon and compatible with iOS?
2/ is the accuracy and speed of the 3.0 release better or at least
similar to the 2.4 release?
3/ is the document layout analysis already stable? A particular need I
have is to be able to get the position of a particular recognized word
in the document? Is this possible with tesseract?
4/ what is the typical preprocessing steps involved in OCR (b&w,
threshold etc.)? Are they already performed by tesseract or do I need
to perform them myself? If yes with which library is it usually done?
Leptonica or OpenCV?

I am also interested if you could give me pointers to code samples
that demonstrate the API usage or tutorials on OCR concepts or on the
APIs of tesseract. Any pointer to the state-of-the-art of OCR,
including papers on useful preprocessing techniques impacting
performance is also welcomed.

I have seen that ScanBizCard is using tesseract 3.0. Do you have other
examples of iPhone applications using Tesseract or concurrent
solutions (commercial or open-source)?

Thanks in advance for all your answers,

Cyril

-- 
You received this message because you are subscribed to the Google
Groups "tesseract-ocr" group.
To post to this group, send email to [email protected]
To unsubscribe from this group, send email to
[email protected]
For more options, visit this group at
http://groups.google.com/group/tesseract-ocr?hl=en

Reply via email to