Hi all,

a quick update on what is happening around DBpedia along the lines of:

1. Freebase
2. Data Quality
3. Live Data 

There were great news from Freebase at ISWC. They are now providing a Linked 
Data interface which makes the complete content of Freebase accessible to the 
Semantic Web.

This is especially exiting for DBpedia as both datasets have a large overlap 
and having RDF access to Freebase makes it easy to mashup and fusion both 
datasets.

We are currently in the process of generating links from DBpedia to Freebase 
for all 2.49 million things in DBpedia. These links will go online sometime 
next week and will immediately allow to mash Freebase and DBpedia data for 
instance using tools like the Marbles Linked Data browser (which does 
owl:sameAs smushing).

The links could also be the foundation for further work on fusing Freebase and 
DBpedia data, which I think will be very exiting and might show that the 
Semantic Web itself is developing into the world's database being fuelled by 
various valuable sources.

There are also good news concerning DBpedia's two main problems: Low data 
quality and stall data.

Georgi, Anja and Paul are getting close to publish a new cleaned-up DBpedia 
dataset based on the current Wikipedia dump. This extraction uses a new 
framework based on manual mappings of hundreds of Wikipedia templates to a 
clean ontology and improved datatype extraction algorithms.  

The new dataset is supposed to be released next week and should be clean enough 
to allow RDFS subsumption reasoning as well as to use it within facet 
browsering UIs.

There is also great progress towards getting the DBpedia dataset current and 
synchronise it with Wikipedia changes: Sören managed to convince the Wikipedia 
foundation to give us access to the Wikipedia live update stream, which tracks 
all changes in Wikipedia itself. Thanks a lot to the foundation for this! This 
is exactly what we needed. Based on this update stream we can sync DBpedia and 
Wikipedia, which will mean about 20 000 updates to the DBpedia dataset per day. 
Orri from OpenLink meant that this is no problem for the Virtuoso server which 
is used to host the DBpedia SPARQL endpoint and Linked Data interface.

Thus after the new dataset is released, we will look into extending the 
extraction framework for continuous updates and are looking forward to be able 
to server a live version a DBpedia soon.

Cheers

Chris
 

--
Chris Bizer
Freie Universität Berlin
Phone: +49 30 838 55509
Mail: [EMAIL PROTECTED]
Web: http://www.bizer.de




-------------------------------------------------------------------------
This SF.Net email is sponsored by the Moblin Your Move Developer's challenge
Build the coolest Linux based applications with Moblin SDK & win great prizes
Grand prize is a trip for two to an Open Source event anywhere in the world
http://moblin-contest.org/redirect.php?banner_id=100&url=/
_______________________________________________
Dbpedia-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

Reply via email to