We did it using scala xml with spark
We start by creating a rdd containing each page is store as a single line :
- split the xml dump with xml_split
- process each split with a shell script which remove "xml_split" tag
and siteinfo section, and put each page on a single line.
- copy resulting files on hdfs
Then the dataset may be load as a text file and processed
val rawDataset = sparkContext.textFile(input)
val allDocuments = rawDataset.map{
case document =>
val page = scala.xml.XML.loadString(document)
val pageTitle = (page \ "title").text
[...]
}
We create a demo using the dataset here: http://wikinsights.org
Le 26/04/2014 23:20, Ajay Nair a écrit :
Is there a way in spark to parse wikipedia xml dump? It seems like the
freebase dump is longer available. Also does the spark shell support the
xml load file sax parser that is present in scala.
Thanks
AJ
Geoffroy Fouquier
http://eXenSa.com