Hi, I have some basic questions before starting a project of OCR recognition for the iPhone.
I have seen the steps to cross-compile tesseract for iOS but have some questions on tesseract roadmap itself: 1/ should I start on tesseract 2.4 or 3.0? From my understanding 3.0 is not yet stable but has a major refactoring ongoing plus several features (including document layout analysis)? The current 3.0 "release" is quite far from the head of the trunk, which do not seem to compile on iOS, so I am wondering if there is any new release (3.01?) planned soon and compatible with iOS? 2/ is the accuracy and speed of the 3.0 release better or at least similar to the 2.4 release? 3/ is the document layout analysis already stable? A particular need I have is to be able to get the position of a particular recognized word in the document? Is this possible with tesseract? 4/ what is the typical preprocessing steps involved in OCR (b&w, threshold etc.)? Are they already performed by tesseract or do I need to perform them myself? If yes with which library is it usually done? Leptonica or OpenCV? I am also interested if you could give me pointers to code samples that demonstrate the API usage or tutorials on OCR concepts or on the APIs of tesseract. Any pointer to the state-of-the-art of OCR, including papers on useful preprocessing techniques impacting performance is also welcomed. I have seen that ScanBizCard is using tesseract 3.0. Do you have other examples of iPhone applications using Tesseract or concurrent solutions (commercial or open-source)? Thanks in advance for all your answers, Cyril -- You received this message because you are subscribed to the Google Groups "tesseract-ocr" group. To post to this group, send email to [email protected] To unsubscribe from this group, send email to [email protected] For more options, visit this group at http://groups.google.com/group/tesseract-ocr?hl=en

