On 8 August 2013 12:55, Hady elsahar <[email protected]> wrote: > I see also that there's no scala installed on the server > we will need that for the LL code to run
Are you sure? If you use Maven, you can just call a Scala launcher. Maven will download all that's necessary. I rarely install Scala on any of the machines where I run DBpedia code. JC > > > > On Thu, Aug 8, 2013 at 12:49 PM, Hady elsahar <[email protected]> wrote: >> >> Hello Dimitris , >> >> sorry for being idle for some days , i was travelling to Leipzig >> >> the python code is running now , don't know how much would it take >> Hopefully 3 hours as Markus mentioned. >> >> Sebastian also gave me access to the server lgd.aksw.org , i'll Run the >> code Again there to speed up the process . then the LLextraction Code >> >> i just want to install some python modules for the code to run and maybe >> some packages like pip to ease the install ? >> is it okay to download packages and modules ? or is there a followed >> protocol in such cases ? >> >> >> thanks >> Regards >> >> >> >> >> >> On Sat, Aug 3, 2013 at 6:25 PM, Dimitris Kontokostas <[email protected]> >> wrote: >>> >>> Hi Hady, >>> >>> This might be what we were waiting for :) >>> If noone else objects, can you create a turtle dump and re-test / adapt >>> your existing ILL code? >>> Afterwards we can start the mappings process >>> >>> Best, >>> Dimitris >>> >>> >>> ---------- Forwarded message ---------- >>> From: Markus Krötzsch <[email protected]> >>> Date: Sat, Aug 3, 2013 at 4:48 PM >>> Subject: [Wikidata-l] Wikidata RDF export available >>> To: "Discussion list for the Wikidata project." >>> <[email protected]> >>> >>> >>> Hi, >>> >>> I am happy to report that an initial, yet fully functional RDF export for >>> Wikidata is now available. The exports can be created using the >>> wda-export-data.py script of the wda toolkit [1]. This script downloads >>> recent Wikidata database dumps and processes them to create RDF/Turtle >>> files. Various options are available to customize the output (e.g., to >>> export statements but not references, or to export only texts in English and >>> Wolof). The file creation takes a few (about three) hours on my machine >>> depending on what exactly is exported. >>> >>> For your convenience, I have created some example exports based on >>> yesterday's dumps. These can be found at [2]. There are three Turtle files: >>> site links only, labels/descriptions/aliases only, statements only. The >>> fourth file is a preliminary version of the Wikibase ontology that is used >>> in the exports. >>> >>> The export format is based on our earlier proposal [3], but it adds a lot >>> of details that had not been specified there yet (namespaces, references, ID >>> generation, compound datavalue encoding, etc.). Details might still change, >>> of course. We might provide regular dumps at another location once the >>> format is stable. >>> >>> As a side effect of these activities, the wda toolkit [1] is also getting >>> more convenient to use. Creating code for exporting the data into other >>> formats is quite easy. >>> >>> Features and known limitations of the wda RDF export: >>> >>> (1) All current Wikidata datatypes are supported. Commons-media data is >>> correctly exported as URLs (not as strings). >>> >>> (2) One-pass processing. Dumps are processed only once, even though this >>> means that we may not know the types of all properties when we first need >>> them: the script queries wikidata.org to find missing information. This is >>> only relevant when exporting statements. >>> >>> (3) Limited language support. The script uses Wikidata's internal >>> language codes for string literals in RDF. In some cases, this might not be >>> correct. It would be great if somebody could create a mapping from Wikidata >>> language codes to BCP47 language codes (let me know if you think you can do >>> this, and I'll tell you where to put it) >>> >>> (4) Limited site language support. To specify the language of linked wiki >>> sites, the script extracts a language code from the URL of the site. Again, >>> this might not be correct in all cases, and it would be great if somebody >>> had a proper mapping from Wikipedias/Wikivoyages to language codes. >>> >>> (5) Some data excluded. Data that cannot currently be edited is not >>> exported, even if it is found in the dumps. Examples include statement ranks >>> and timezones for time datavalues. I also currently exclude labels and >>> descriptions for simple English, formal German, and informal Dutch, since >>> these would pollute the label space for English, German, and Dutch without >>> adding much benefit (other than possibly for simple English descriptions, I >>> cannot see any case where these languages should ever have different >>> Wikidata texts at all). >>> >>> Feedback is welcome. >>> >>> Cheers, >>> >>> Markus >>> >>> [1] https://github.com/mkroetzsch/wda >>> Run "python wda-export.data.py --help" for usage instructions >>> [2] http://semanticweb.org/RDF/Wikidata/ >>> [3] http://meta.wikimedia.org/wiki/Wikidata/Development/RDF >>> >>> -- >>> Markus Kroetzsch, Departmental Lecturer >>> Department of Computer Science, University of Oxford >>> Room 306, Parks Road, OX1 3QD Oxford, United Kingdom >>> +44 (0)1865 283529 http://korrekt.org/ >>> >>> _______________________________________________ >>> Wikidata-l mailing list >>> [email protected] >>> https://lists.wikimedia.org/mailman/listinfo/wikidata-l >>> >>> >>> >>> -- >>> Kontokostas Dimitris >> >> >> ------------------------------------------------- >> Hady El-Sahar >> Research Assistant >> Center of Informatics Sciences | Nile University >> >> > > > > -- > ------------------------------------------------- > Hady El-Sahar > Research Assistant > Center of Informatics Sciences | Nile University > > ------------------------------------------------------------------------------ Get 100% visibility into Java/.NET code with AppDynamics Lite! It's a free troubleshooting tool designed for production. Get down to code-level detail for bottlenecks, with <2% overhead. Download for free and get started troubleshooting in minutes. http://pubads.g.doubleclick.net/gampad/clk?id=48897031&iu=/4140/ostg.clktrk _______________________________________________ Dbpedia-developers mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/dbpedia-developers
