Hi,

you'd need as well to convert these document to pure text in order to process them; you can try to spawn OOo for conversion in a batch mode but the easier option is to use unzip in a script, and take content.xml only from the files. Then process the files using awk (define the field separator just like you would define the word boundary) and filter out all words that match <[a-z]+>. This should kill all xml from the files.

Regards
Marcin

ge pisze:
Hello,

I did word collection several times using different sources( web sources)

I use linux, but these tools are also available
for windows as gnu tools.

I used awk, like:
 for (i = 1; i <= $NF; i++)
   print $i;

This prints each word in a single line.

Then I sorted the file using sort <infile > outfile
and then used further awk scripts to get rid of word endings,
this is probably much easier for Danish, than for Hungarian.

Good luck! Eleonora


[lingu-dev] Help needed - bulk extraction of words

Hi all,
The Danish project has been so fortunate to receive a bunch of articles from a news magazine. These are odt files and we would like to extract the words from these documents. We have programs for this purpose, but we usually get donations one document at the time. This time we have several thousand documents and I believe it would take about a year to load these documents one by one.

Do any of you have a program that can extract words from several documents ?

The words will be loaded into our workflow for linguistic processing and at the end be a part of the Danish spelling directory.

Thanks in advance.



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to