Hi Dileepa,
some thoughts that I also share with other Marmotta's team members
regarding integration with Open Refine.
For the second level of integration, that fundamentally exports towards
Marmotta both CSV and other data to produce RDF, it would be interesting
try to add the functionality in Open Refine to supply additional data to
dataset, using for example DCAT Vocabulary [1].
I don't remember if this feature is covered by GRefine RDF Extension, of
that it's present a new release(ALPHA 0.9.0) [2]
If dataset is supplied with DCAT metadata, Marmotta could expose it to
facilitate its interoperability with other datasets.
To do that, Marmotta needs to store also structured datasets, not
necessarily instantiated in RDF triples.
What do you think about?
Regards,
Raffaele.


[1] http://www.w3.org/TR/vocab-dcat/
[2] https://github.com/fadmaa/grefine-rdf-extension/releases/tag/v0.9.0


On 11 March 2014 10:29, Dileepa Jayakody <[email protected]> wrote:

> Thank you very much Raffaele for the detailed explaination.
>
> I will do some more background research on Marmotta data import and
> OpenRefine and come up with questions and ideas I get.
>
> Also any new suggestions, directions to evolve this project idea are
> welcome.
>
> Thanks,
> Dileepa
>
>
> On Tue, Mar 11, 2014 at 3:14 AM, Raffaele Palmieri <
> [email protected]> wrote:
>
> > Hi Dileepa,
> > pleased to meet you and know your interest for contributing to Marmotta.
> > As discussed in Marmotta's mailing list, this integration could be
> reached
> > at various levels.
> > A first level is reached refining your messy data with Refine tools,
> using
> > RDF extension, that already offers a graphical UI to model RDF data
> > producing an RDF skeleton and then import new data in Marmotta, compliant
> > to the created skeleton .
> > This integration mode has been implemented in the past using [1] but
> needs
> > to be updated because:
> > 1)Google Refine became Open Refine
> > 2)LMF became Marmotta in its linked-data core functionalities
> > This update also requires work about project configuration, because Open
> > Refine has a different configuration than Apache Marmotta.
> > Whatever kind of integration could be achieved, I think that work about
> > project configuration is required.
> > A second level of integration is reached if you break up RDF in CSV and
> set
> > of RDF mappings(aka RDF skeleton).
> > So, starting from exported project that contains CSV and related actions
> to
> > produce RDF Skeleton, the integration expects to produce final RDF in
> > Marmotta's world, probably performing similar steps as GRefine RDF
> > Extension.
> > For that second level of integration, export functionality and RDF
> skeleton
> > should be explored to verify what is easily exportable.
> > At the moment, these are the hypothesis of integration, clearly the
> second
> > appears to be more complex, but also the first brings non-trivial work.
> > Since you have experience on other projects related to Semantic Web, as
> > Apache Stanbol, feel free to propose other hypothesis of integration,
> > regards,
> > Raffaele.
> >
> > [1]https://code.google.com/p/lmf/wiki/GoogleRefineExtension
> >
> >
> >
> >
> > On 10 March 2014 21:35, Dileepa Jayakody <[email protected]>
> > wrote:
> >
> > > Hi All,
> > >
> > > I'm Dileepa a research student from University of Moratuwa, Sri Lanka
> > with
> > > keen interest in the linked-data and semantic-web domains. I have
> worked
> > > with linked-data related projects such as Apache Stanbol and I'm
> > > experienced with related technologies like RDF, SPARQL, FOAF etc. I'm
> > very
> > > much interested in applying for GSoC this year with Apache Marmotta.
> > >
> > > I would like to open up a discussion on the OpenRefine integration
> > project
> > > idea [1]. AFAIU, the goal of this project is to import data to Marmotta
> > > triple store (to Kiwi triple-store by default) from OpenRefine after
> the
> > > data has been refined and exported.
> > >
> > > I did some background reading on Marmotta data import process [2] which
> > > explains different ways to import RDF data to back-end triple store.
> > > Currently OpenRefine exports data in several formats: csv, tsv, xsl,
> html
> > > tables. So I think the main task of this project will be to convert
> this
> > > exported data into RDF format and make it compatible to Marmotta data
> > > import process. I did a quick research on how to do so and there are
> > > several options to convert such data to RDF.
> > >
> > > They are,
> > > 1. RDF extension to OpenRefine :
> > https://github.com/sparkica/rdf-extension
> > > 2. RDF refine : http://refine.deri.ie/
> > > 3. D2R server http://d2rq.org/d2r-server (if OpenRefine data is
> imported
> > > from a SQL database)
> > >
> > > Apart from the data conversion process from OpenRefine to RDF, what are
> > the
> > > other tasks to be done in this project?
> > > Appreciate your thoughts and suggestions.
> > >
> > > Thanks,
> > > Dileepa
> > >
> > > [1] https://issues.apache.org/jira/browse/MARMOTTA-202
> > > [2] http://wiki.apache.org/marmotta/ImportData
> > > [3]
> > >
> >
> https://github.com/OpenRefine/OpenRefine/wiki/Exporters#exporting-projects
> > >
> >
>

Reply via email to