Thank you very much Raffaele for the detailed explaination.

I will do some more background research on Marmotta data import and
OpenRefine and come up with questions and ideas I get.

Also any new suggestions, directions to evolve this project idea are
welcome.

Thanks,
Dileepa


On Tue, Mar 11, 2014 at 3:14 AM, Raffaele Palmieri <
[email protected]> wrote:

> Hi Dileepa,
> pleased to meet you and know your interest for contributing to Marmotta.
> As discussed in Marmotta's mailing list, this integration could be reached
> at various levels.
> A first level is reached refining your messy data with Refine tools, using
> RDF extension, that already offers a graphical UI to model RDF data
> producing an RDF skeleton and then import new data in Marmotta, compliant
> to the created skeleton .
> This integration mode has been implemented in the past using [1] but needs
> to be updated because:
> 1)Google Refine became Open Refine
> 2)LMF became Marmotta in its linked-data core functionalities
> This update also requires work about project configuration, because Open
> Refine has a different configuration than Apache Marmotta.
> Whatever kind of integration could be achieved, I think that work about
> project configuration is required.
> A second level of integration is reached if you break up RDF in CSV and set
> of RDF mappings(aka RDF skeleton).
> So, starting from exported project that contains CSV and related actions to
> produce RDF Skeleton, the integration expects to produce final RDF in
> Marmotta's world, probably performing similar steps as GRefine RDF
> Extension.
> For that second level of integration, export functionality and RDF skeleton
> should be explored to verify what is easily exportable.
> At the moment, these are the hypothesis of integration, clearly the second
> appears to be more complex, but also the first brings non-trivial work.
> Since you have experience on other projects related to Semantic Web, as
> Apache Stanbol, feel free to propose other hypothesis of integration,
> regards,
> Raffaele.
>
> [1]https://code.google.com/p/lmf/wiki/GoogleRefineExtension
>
>
>
>
> On 10 March 2014 21:35, Dileepa Jayakody <[email protected]>
> wrote:
>
> > Hi All,
> >
> > I'm Dileepa a research student from University of Moratuwa, Sri Lanka
> with
> > keen interest in the linked-data and semantic-web domains. I have worked
> > with linked-data related projects such as Apache Stanbol and I'm
> > experienced with related technologies like RDF, SPARQL, FOAF etc. I'm
> very
> > much interested in applying for GSoC this year with Apache Marmotta.
> >
> > I would like to open up a discussion on the OpenRefine integration
> project
> > idea [1]. AFAIU, the goal of this project is to import data to Marmotta
> > triple store (to Kiwi triple-store by default) from OpenRefine after the
> > data has been refined and exported.
> >
> > I did some background reading on Marmotta data import process [2] which
> > explains different ways to import RDF data to back-end triple store.
> > Currently OpenRefine exports data in several formats: csv, tsv, xsl, html
> > tables. So I think the main task of this project will be to convert this
> > exported data into RDF format and make it compatible to Marmotta data
> > import process. I did a quick research on how to do so and there are
> > several options to convert such data to RDF.
> >
> > They are,
> > 1. RDF extension to OpenRefine :
> https://github.com/sparkica/rdf-extension
> > 2. RDF refine : http://refine.deri.ie/
> > 3. D2R server http://d2rq.org/d2r-server (if OpenRefine data is imported
> > from a SQL database)
> >
> > Apart from the data conversion process from OpenRefine to RDF, what are
> the
> > other tasks to be done in this project?
> > Appreciate your thoughts and suggestions.
> >
> > Thanks,
> > Dileepa
> >
> > [1] https://issues.apache.org/jira/browse/MARMOTTA-202
> > [2] http://wiki.apache.org/marmotta/ImportData
> > [3]
> >
> https://github.com/OpenRefine/OpenRefine/wiki/Exporters#exporting-projects
> >
>

Reply via email to