Hi everyone,
I'm trying to use the Information Extraction Framework, but i should be
doing something wrong and I'm having problems with the dumps.
I download the dump "eswikisource-20100317-pages-articles.xml.bz2" I saved
it in a folder, I setup the configuration dumpDir to the folder and I try to
run the extraction..but...
[INFO] launcher 'Extract' selected => org.dbpedia.extraction.Extract
Exception in thread "Thread-1" java.lang.Exception: Dump directory not
found: /home/rober/Escritorio/dbpedia/datos/pages/commons
at
org.dbpedia.extraction.ConfigLoader$Config.getDumpFile(ConfigLoader.scala:93)
at
org.dbpedia.extraction.ConfigLoader$Config.<init>(ConfigLoader.scala:85)
at org.dbpedia.extraction.ConfigLoader$.load(ConfigLoader.scala:28)
at org.dbpedia.extraction.Extract$ExtractionThread.run(Extract.scala:26)
[INFO]
------------------------------------------------------------------------
[INFO] BUILD SUCCESSFUL
Reading the doc I saw this "The dump files should be organized in the way as
they are on the wikipedia servers.
e.g. {dumpDir}/sc/20100306/scwiki-20100306-pages-articles.xml.bz2. In
addition to the dumps of the configured languages, you'll need the Wikipedia
Commons Dump."
Now I'm not sure what is "the Wikipedia Commons Dump"... or if I'm using a
wrong dump
Can anyone help me?
Thanks for the attention.
------------------------------------------------------------------------------
Download Intel® Parallel Studio Eval
Try the new software tools for yourself. Speed compiling, find bugs
proactively, and fine-tune applications for parallel performance.
See why Intel Parallel Studio got high marks during beta.
http://p.sf.net/sfu/intel-sw-dev
_______________________________________________
Dbpedia-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion