Hi,

 How can i merge all this files in order to have ALL informations about
one wikipedia-page in the same file ?


Are you using allmighty unix? Then you would have all the tools at hand, in order to do a hash based split based on the subjects of the triples.

with "bzcat *" you could merge ALL files (in a directory) into a single stream.
(If you just want everything in a single huge file, youre done now)

Then you pipe the stream into a little script something like:
#split each line into subject and remainder
|while read subject rest; do

    h=compute hash based on subject

    # use the hash as a filename, and append our current line to that file
    echo "$subject $rest" >> $h
done
|
This would cause all triples with same subject to end up in the same file.

Afterwards you could use "sort -u filename" to sort the lines in a file, which would result in all triples with the same subject to be in consecutive rows.
(the -u would remove potential duplicate rows)

Kind regards,
Claus


On 01/13/2011 05:40 PM, Julien wrote:
Hi,

Thx for this reply

On Thu, Jan 13, 2011 at 5:26 PM, Max Jakob<[email protected]>  wrote:
Hi Julien,

On Thu, Jan 13, 2011 at 16:38, Julien<[email protected]>  wrote:
Hi,
For a school project, i've to use semantic data. So i would like to use
DBPedia.
But, the available datasets on this page
(http://wiki.dbpedia.org/Downloads36) are old, and in a strange format. I
don't manage to use this data...
Well, this data was extracted using the most recent English Wikipedia
dump (http://download.wikimedia.org/enwiki/) and using other language
Wikipedia dumps of around the same time.

Ok.

Can i download all dbpedia datasets in json ? Like in this page, for example
:http://dbpedia.org/data/The_Red_and_the_Black.json
Or in rdf ? xml ?
The files are bzip2 compressed N-Triples, a serialization of RDF. From
that there are straightforward ways to transform the data into JSON or
XML.


Yes, but there is one file for Ontology, one file for Titles, another
for abstracts... etc...
How can i merge all this files in order to have ALL informations about
one wikipedia-page in the same file ?

In this page :http://dbpedia.org/data/The_Red_and_the_Black.json
all data about this entity are in this page. How can i do this ?


Hope this helps.

Cheers,
Max
------------------------------------------------------------------------------
Protect Your Site and Customers from Malware Attacks
Learn about various malware tactics and how to avoid them. Understand
malware threats, the impact they can have on your business, and how you
can protect your company and customers by using code signing.
http://p.sf.net/sfu/oracle-sfdevnl
_______________________________________________
Dbpedia-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

------------------------------------------------------------------------------
Protect Your Site and Customers from Malware Attacks
Learn about various malware tactics and how to avoid them. Understand 
malware threats, the impact they can have on your business, and how you 
can protect your company and customers by using code signing.
http://p.sf.net/sfu/oracle-sfdevnl
_______________________________________________
Dbpedia-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

Reply via email to