Hi, 

i'm responsible for the wiktionary extraction.

As Sebastian said, dbpedia will give you the categorization and
infoboxes (you will need to configure the mappings from property names
to URL's etc but i guess wikitravel is not so complex?). That should be
easy and produce quite some data. If you want to go deeper, the generic
extractor we build for wiktionary, could help. It uses templates
containing placeholders, that extract data in a declarative way. It
could be usefull for lists of data like the restaurants and hotels
section. But Infoboxes are not parseable, so this is just a addition. I
could help a bit with the config if you have specific questions. Do you
only want English?

Regards,
Jonas

Am Donnerstag, den 01.03.2012, 17:07 +0100 schrieb Sebastian Hellmann:
> Hi Amulya,
> here is my educated guess: Yes, there are quite a few structure, which
> you can extract with the DBpedia framework.
> You will get infoboxes and some other data. It is definitely worth to
> try it. 
> 
> We are currently working on extensions for other Wikis, especially
> converting Wiktionary to RDF with a generic extractor.
> Please look here:
> http://dbpedia.hg.sourceforge.net/hgweb/dbpedia/extraction_framework/file/f4da35d167ce/wiktionary/config-en.xml
> http://dbpedia.hg.sourceforge.net/hgweb/dbpedia/extraction_framework/file/f4da35d167ce/wiktionary/config-de.xml
> 
> These are xml configurations for the German and the English
> Wiktionary.
> 
> I also cc'ed Jonas who develops the plugin.
> We would be happy to provide downloads for the data you produce ....
> 
> Tell me if you need sourceforge access to branch the mercurial repo.
> All the best,
> Sebastian
> 
> On 02/29/2012 08:56 AM, amulya rattan wrote: 
> > Hi all,
> > 
> > Inspired by what dbpedia does with wikipedia, I wanted to try creating a
> > similar dataset for wikitravel.org. Before going through how dbpedia does
> > the same for wikipedia, I wanted to know if the solution used by dbpedia
> > could be easily transferrable to this project or a completely fresh
> > approach should be used. Looks like the dataset created for this would be
> > much smaller than for dbpedia, thus I want to make sure I use the right
> > approach. All opinions and inputs are highly appreciated.
> > 
> > Thanks,
> > ~Amulya
> > 
> > 
> > ------------------------------------------------------------------------------
> > Virtualization & Cloud Management Using Capacity Planning
> > Cloud computing makes use of virtualization - but cloud computing 
> > also focuses on allowing computing to be delivered as a service.
> > http://www.accelacomm.com/jaw/sfnl/114/51521223/
> > 
> > _______________________________________________
> > Dbpedia-discussion mailing list
> > [email protected]
> > https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion
> 
> 
> -- 
> Dipl. Inf. Sebastian Hellmann
> Department of Computer Science, University of Leipzig
> Projects: http://nlp2rdf.org , http://dbpedia.org
> Homepage: http://bis.informatik.uni-leipzig.de/SebastianHellmann
> Research Group: http://aksw.org




------------------------------------------------------------------------------
Virtualization & Cloud Management Using Capacity Planning
Cloud computing makes use of virtualization - but cloud computing 
also focuses on allowing computing to be delivered as a service.
http://www.accelacomm.com/jaw/sfnl/114/51521223/
_______________________________________________
Dbpedia-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

Reply via email to