There are two separate issues here - HTML parsing (sometimes called cleanup) vs. getting rid of boilerplate content, which is also often called HTML cleanup.

TagSoup & NekoHTML are examples of the former - code that "fixes up" HTML documents so you can apply standard XML parsing techniques.

The articles originally referenced below, as well as my prior note about nCleaner, are talking about the latter - trying to get rid of headers, footers, ads, etc.

-- Ken

On Nov 28, 2009, at 12:30pm, Marc Hofer wrote:

Hi Drew,

currently we are using a HTML Filter module of the Univeristy Duisburg-Essen, that can be found here: http://www.is.informatik.uni-duisburg.de/projects/java-unidu/filter.html

Another idea was to try Jericho or NekoHTML.
http://www.java2s.com/Product/Java/Development/HTML-Parser.htm

Thanks for your advice, we will test it and let you know, whether it works well.

Marc

Drew Farris schrieb:
Hi Marc,
How are you planning on cleaning up the HTML documents?
Perhaps something like this would be useful: I came across an
interesting approach a few days ago, it would be interesting to hear
more from someone who has tried something like this:
http://ai-depot.com/articles/the-easy-way-to-extract-useful-text-from-arbitrary-html/
Described further, with java implementations here:
http://sujitpal.blogspot.com/2009/11/extracting-useful-text-from-html.html
Drew
On Sat, Nov 28, 2009 at 2:57 PM, Marc Hofer <m...@marc-hofer.de> wrote:
Hello everybody,

having already presented the draft of our architecture, I would like now to discuss the second layer more in detail. As mentioned before we have chosen UIMA for this layer. The main aggregate currently consists of the Whitespace Tokenizer Annotator, the Snowball Annotator (Stemming) and a list- based StopwordFilter. Before processing this aggregate in a map-only job in Hadoop, we want to filter all HTML tags and forward only this preprocessed data to the aggregate. The reason for this is that it is difficult to change the document during processing in UIMA and it is impractical to work all the
time on documents containing HTML tags.

Furthermore we are planning to add the Tagger Annotator, which implements a Hidden Markov Model tagger. Here we aren't sure, which tokens with their corresponding part of speech tags to delete or not and so using them for the feature extraction. One purpose could be to use at the very beginning only
substantives and verbs.

We are very interested in your comments and remarks and it would be nice to
hear from you.

Cheers,
Marc



--------------------------------------------
Ken Krugler
+1 530-210-6378
http://bixolabs.com
e l a s t i c   w e b   m i n i n g




Reply via email to