Hi Dileepa,
pleased to meet you and know your interest for contributing to Marmotta.
As discussed in Marmotta's mailing list, this integration could be reached
at various levels.
A first level is reached refining your messy data with Refine tools, using
RDF extension, that already offers a graphical UI to model RDF data
producing an RDF skeleton and then import new data in Marmotta, compliant
to the created skeleton .
This integration mode has been implemented in the past using [1] but needs
to be updated because:
1)Google Refine became Open Refine
2)LMF became Marmotta in its linked-data core functionalities
This update also requires work about project configuration, because Open
Refine has a different configuration than Apache Marmotta.
Whatever kind of integration could be achieved, I think that work about
project configuration is required.
A second level of integration is reached if you break up RDF in CSV and set
of RDF mappings(aka RDF skeleton).
So, starting from exported project that contains CSV and related actions to
produce RDF Skeleton, the integration expects to produce final RDF in
Marmotta's world, probably performing similar steps as GRefine RDF
Extension.
For that second level of integration, export functionality and RDF skeleton
should be explored to verify what is easily exportable.
At the moment, these are the hypothesis of integration, clearly the second
appears to be more complex, but also the first brings non-trivial work.
Since you have experience on other projects related to Semantic Web, as
Apache Stanbol, feel free to propose other hypothesis of integration,
regards,
Raffaele.

[1]https://code.google.com/p/lmf/wiki/GoogleRefineExtension




On 10 March 2014 21:35, Dileepa Jayakody <[email protected]> wrote:

> Hi All,
>
> I'm Dileepa a research student from University of Moratuwa, Sri Lanka with
> keen interest in the linked-data and semantic-web domains. I have worked
> with linked-data related projects such as Apache Stanbol and I'm
> experienced with related technologies like RDF, SPARQL, FOAF etc. I'm very
> much interested in applying for GSoC this year with Apache Marmotta.
>
> I would like to open up a discussion on the OpenRefine integration project
> idea [1]. AFAIU, the goal of this project is to import data to Marmotta
> triple store (to Kiwi triple-store by default) from OpenRefine after the
> data has been refined and exported.
>
> I did some background reading on Marmotta data import process [2] which
> explains different ways to import RDF data to back-end triple store.
> Currently OpenRefine exports data in several formats: csv, tsv, xsl, html
> tables. So I think the main task of this project will be to convert this
> exported data into RDF format and make it compatible to Marmotta data
> import process. I did a quick research on how to do so and there are
> several options to convert such data to RDF.
>
> They are,
> 1. RDF extension to OpenRefine : https://github.com/sparkica/rdf-extension
> 2. RDF refine : http://refine.deri.ie/
> 3. D2R server http://d2rq.org/d2r-server (if OpenRefine data is imported
> from a SQL database)
>
> Apart from the data conversion process from OpenRefine to RDF, what are the
> other tasks to be done in this project?
> Appreciate your thoughts and suggestions.
>
> Thanks,
> Dileepa
>
> [1] https://issues.apache.org/jira/browse/MARMOTTA-202
> [2] http://wiki.apache.org/marmotta/ImportData
> [3]
> https://github.com/OpenRefine/OpenRefine/wiki/Exporters#exporting-projects
>

Reply via email to