Thank you very much Raffaele for the detailed explaination. I will do some more background research on Marmotta data import and OpenRefine and come up with questions and ideas I get.
Also any new suggestions, directions to evolve this project idea are welcome. Thanks, Dileepa On Tue, Mar 11, 2014 at 3:14 AM, Raffaele Palmieri < [email protected]> wrote: > Hi Dileepa, > pleased to meet you and know your interest for contributing to Marmotta. > As discussed in Marmotta's mailing list, this integration could be reached > at various levels. > A first level is reached refining your messy data with Refine tools, using > RDF extension, that already offers a graphical UI to model RDF data > producing an RDF skeleton and then import new data in Marmotta, compliant > to the created skeleton . > This integration mode has been implemented in the past using [1] but needs > to be updated because: > 1)Google Refine became Open Refine > 2)LMF became Marmotta in its linked-data core functionalities > This update also requires work about project configuration, because Open > Refine has a different configuration than Apache Marmotta. > Whatever kind of integration could be achieved, I think that work about > project configuration is required. > A second level of integration is reached if you break up RDF in CSV and set > of RDF mappings(aka RDF skeleton). > So, starting from exported project that contains CSV and related actions to > produce RDF Skeleton, the integration expects to produce final RDF in > Marmotta's world, probably performing similar steps as GRefine RDF > Extension. > For that second level of integration, export functionality and RDF skeleton > should be explored to verify what is easily exportable. > At the moment, these are the hypothesis of integration, clearly the second > appears to be more complex, but also the first brings non-trivial work. > Since you have experience on other projects related to Semantic Web, as > Apache Stanbol, feel free to propose other hypothesis of integration, > regards, > Raffaele. > > [1]https://code.google.com/p/lmf/wiki/GoogleRefineExtension > > > > > On 10 March 2014 21:35, Dileepa Jayakody <[email protected]> > wrote: > > > Hi All, > > > > I'm Dileepa a research student from University of Moratuwa, Sri Lanka > with > > keen interest in the linked-data and semantic-web domains. I have worked > > with linked-data related projects such as Apache Stanbol and I'm > > experienced with related technologies like RDF, SPARQL, FOAF etc. I'm > very > > much interested in applying for GSoC this year with Apache Marmotta. > > > > I would like to open up a discussion on the OpenRefine integration > project > > idea [1]. AFAIU, the goal of this project is to import data to Marmotta > > triple store (to Kiwi triple-store by default) from OpenRefine after the > > data has been refined and exported. > > > > I did some background reading on Marmotta data import process [2] which > > explains different ways to import RDF data to back-end triple store. > > Currently OpenRefine exports data in several formats: csv, tsv, xsl, html > > tables. So I think the main task of this project will be to convert this > > exported data into RDF format and make it compatible to Marmotta data > > import process. I did a quick research on how to do so and there are > > several options to convert such data to RDF. > > > > They are, > > 1. RDF extension to OpenRefine : > https://github.com/sparkica/rdf-extension > > 2. RDF refine : http://refine.deri.ie/ > > 3. D2R server http://d2rq.org/d2r-server (if OpenRefine data is imported > > from a SQL database) > > > > Apart from the data conversion process from OpenRefine to RDF, what are > the > > other tasks to be done in this project? > > Appreciate your thoughts and suggestions. > > > > Thanks, > > Dileepa > > > > [1] https://issues.apache.org/jira/browse/MARMOTTA-202 > > [2] http://wiki.apache.org/marmotta/ImportData > > [3] > > > https://github.com/OpenRefine/OpenRefine/wiki/Exporters#exporting-projects > > >
