[
https://issues.apache.org/jira/browse/TIKA-2038?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15399886#comment-15399886
]
Tim Allison edited comment on TIKA-2038 at 7/29/16 7:26 PM:
------------------------------------------------------------
I'm attaching the raw results from running Tika against the corpus available on
IUST-HTMLCharDet's github
[site|https://github.com/shabanali-faghani/IUST-HTMLCharDet/tree/master/test-data/encoding-wise].
was (Author: [email protected]):
I'm attaching the raw results from running Tika against the corpus available on
UST-HTMLCharDet's github
[site|https://github.com/shabanali-faghani/IUST-HTMLCharDet/tree/master/test-data/encoding-wise].
> A more accurate facility for detecting Charset Encoding of HTML documents
> -------------------------------------------------------------------------
>
> Key: TIKA-2038
> URL: https://issues.apache.org/jira/browse/TIKA-2038
> Project: Tika
> Issue Type: Improvement
> Components: core, detector
> Reporter: Shabanali Faghani
> Priority: Minor
> Attachments: tika_1_14-SNAPSHOT_encoding_detector.zip
>
>
> Currently, Tika uses icu4j for detecting charset encoding of HTML documents
> as well as the other naturally text documents. But the accuracy of encoding
> detector tools, including icu4j, in dealing with the HTML documents is
> meaningfully less than from which the other text documents. Hence, in our
> project I developed a library that works pretty well for HTML documents,
> which is available here: https://github.com/shabanali-faghani/IUST-HTMLCharDet
> Since Tika is widely used with and within some of other Apache stuffs such as
> Nutch, Lucene, Solr, etc. and these projects are strongly in connection with
> the HTML documents, it seems that having such an facility in Tika also will
> help them to become more accurate.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)