I am exploring ways to improve the performance(single-thread / 
multi-thread) of tesseract-ocr inference. Performance profiling tess-ocr 
inference (CLI) using Intel Vtune 
<https://www.intel.com/content/www/us/en/developer/tools/oneapi/vtune-profiler.html>
 
showed very little usage of the CPU's vector registers (AVX-2 and 512). 
The default inference using tesseract works on a single image 
(batch_size=1). I think batched inference would drive better utilization of 
the vector registers and thus improve the inference throughput of tess-ocr. 
Is there a way to enable batched inference on tess-ocr ?

Thanks in advance

Regards
Vishnu


other details:
- system : Intel Xeon 8380 (Icelake), Ubuntu 22.04(5.15), GCC 11.3
- tesseract : built from source (5.3.0), tessdata
- configure flags : --enable-float32 --disable-opencl --disable-graphics 
'CXXFLAGS=-O3 -mavx512f -mfma'

other performance observations:
- considerable openmp pause times observed for the multi threaded runs
- LSTM --> Tanh lookup times are considerably high (~ 50ms per lookup)

-- 
You received this message because you are subscribed to the Google Groups 
"tesseract-ocr" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/tesseract-ocr/982ae47a-6b2f-4565-9e82-0aadf1f09d65n%40googlegroups.com.

Reply via email to