If you're a Real Geek you know this is a great time of the year to play
around with new data toys! :)
so give Kibi 0.2 a spin, for it is finally here...
*What is Kibi?*
Kibi 0.2 is a Kibana fork for Data Intelligence.
For those that dont know Kibana, it's a quite amazing data search/analytics
Hi Sebastian, just for context
(i am collaborating with a leadingmarket data provider) there are 17 M+
organizations in italy alone (either alive or dead .. but maybe worth still
being in a database).
Maaaybe, just maaybe its worth tto talk to some of these organization and
campaign the opening
://siren.solutions/kibi/docs/current/
Gio
On Mon, Sep 21, 2015 at 3:30 PM, Alfredo Serafini <ser...@gmail.com> wrote:
> wow, it seems a great work! :-)
>
> I'll try it as soon as I can, thank you for sharing
>
> 2015-09-21 15:56 GMT+02:00 Giovanni Tummarello <g.tummare...@g
partner's media. Acknowledgement also
go to the Data and Knowledge Management Research Unit at Fondazione Bruno
Kessler FBK institute for the project idea support and feedback.
Giovanni Tummarello
SIREn Solutions (Formerly Sindicetech)
Hi Dimistris, everyone in the team. congratulations, great job.. it will
certainly be useful
Gio
On Fri, May 15, 2015 at 11:28 AM, Dimitris Kontokostas
kontokos...@informatik.uni-leipzig.de wrote:
Dear all,
Following up on the early prototype we announced earlier [1] we are happy
to
Hi Hugh, Christian
not sure. This[1] takes RDF as an input, the user couldnt care less about
the internal, just gets a fast, powerful relational browser. Nothing here
is SPARQL there isnt even a triplestore. RDF is processed via hadop, Solr
indexes are created, no URI is looked up.
Still it
Hugh,
i think if you send them down a route where you have to write bespoke
software (which uses RDF concept, hard to find developers to write and
maintain) for purposes for which mature widely tested and widely spread
software exists you'd be doing them a disservice.
Eventually they'll find
Hi all,
would anyone be able to recommend/point to available system/library to get
opengraph search capabilities like.. like in facebook .. things like
pictures that my friends like
people depicted in pictures posted by my friends
thanks in advance.
Gio
Great job! Clearly when it rains, it pours!
Lot's of great Linked Data visualizations are now popping up everywhere,
just what we all needed.
hi kingsley
which other are you referring to? (i likely have missed them)
thanks
Gio
Hi Chris, this is interesting, and its great you're looking also at the
world of marked up data.
my 2c shortly
If you build an application that requires
DBpedia/YAGO/Freebase/UMBEL/Cyc-style general knowledge about entities, or
you build an applications that requires geographic, live
Chris hi,
i would be interested in discussing what is the message that will
accompany this new version?
If i am not wrong there appear to be more bubbles than last time here so
i wonder is the message that's going out with this diagram that adoption
has increased (e.g. as there were 200 and
Dear all,
a Ph.D position in the Web of Data Unit, University of Trento/ FBK
Institute is now accepting submissions with deadline 16/6.
This position is co-sponsored by www.spaziodati.eu which will also make
available to the candidate highly valuable, not otherwise available data as
provided by
Dear all,
we're glad to say the article, also in improved version, is now available
on SemanticWeb.com
http://semanticweb.com/end-support-sindice-com-search-engine-history-lessons-learned-legacy-guest-post_b42797
cheers
Gio
On Wed, Apr 30, 2014 at 3:05 AM, Giovanni Tummarello g.tummare
and as a way to glimpse at future directions of this
field and technologies.
The Sindice.com Founders
Dr. Giovanni Tummarello Dr. Renaud Delbru
http://blog.sindice.com/2014/04/28/end-of-support-for-
sindice-com-history-and-legacy/
[1]
http://webcache.googleusercontent.com/search?q=cache:09H7ZzKW8AcJ:blog.sindice.com/2014/04/28/end-of-support-for-sindice-com-history-and-legacy/+cd=1hl=enct=clnk
On Tue, Apr 29, 2014 at 6:28 PM, Giovanni Tummarello g.tummare...@gmail.com
wrote:
Dear all,
the Sindice team announces today
Hi all
also with the very kind support of the RDF guys at google dev relationship
(thanks Dan! :)
SindiceTech is happy to announce the availability of our Freebase
Distribution in the clouds.
*A Freebase data distribution that's easy to use.*
Freebase is an amazing data resource at the core of
.
On 28 Jan 2014, at 14:19, Giovanni Tummarello g.tummare...@gmail.com
wrote:
With respect to Sindice
for a number of reasons, the people who originally created it, the
former Data Intensive Infrastructure group, are either not working in the
original institution hosting it, National
With respect to Sindice
for a number of reasons, the people who originally created it, the former
Data Intensive Infrastructure group, are either not working in the original
institution hosting it, National University of Ireland Galway, institute
formerly known as DERI or have been assigned to
Science fiction (ooh ooh ooh) double feature
Doctor X (ooh ooh ooh) will build a creature
See androids fighting (ooh ooh ooh) Brad and Janet
Anne Francis stars in (ooh ooh ooh) Forbidden Planet
Wo oh oh oh oh oh
At the late night, double feature, picture show
http://www.deezer.com/album/238003
-- Forwarded message --
From: Lewis John Mcgibbney lewis.mcgibb...@gmail.com
Date: Wed, Jul 10, 2013 at 11:06 PM
Subject: [ANNOUNCEMENT] Apache Any23 0.8.0 Release
To: u...@any23.apache.org, d...@any23.apache.org d...@any23.apache.org
Hi All,
The Any23 PMC are pleased to
My 2c is .. i agree with kingsley diagram , linked data should be possible
without RDF (no matter serialization) :)
however this is different from previous definitions
i think its a step forward.. but it is different from previously. Do we
want to call it Linked Data 2.0? under this definition
:54 AM, Giovanni Tummarello wrote:
My 2c is .. i agree with kingsley diagram , linked data should be
possible without RDF (no matter serialization) :)
however this is different from previous definitions
i think its a step forward.. but it is different from previously. Do we
want to call
Interesting kingsley, not sure what the implication is of GPL2 is e.g.
would one have to redistribute the whole source code of anything attached
to it?
anyway great,
Gio
On Tue, Jun 11, 2013 at 6:13 PM, Kingsley Idehen kide...@openlinksw.comwrote:
All,
Here is a link [1][2] demonstrating
Completely agree Hugh, lets make sure we stick to the thread.
Gio
On Thu, Apr 18, 2013 at 11:41 AM, Hugh Glaser h...@ecs.soton.ac.uk wrote:
Someone starts a thread (in this case Luca and his Restpark), about something
they would like to get some feedback on.
In the very first reply, an issue
Anyone seriously sparql/java/javascript interested in a 4 month
project to bring a fundamentally useful semweb tool to its full open
source potential?
well rewarded, visit Ireland, work with the Sindice team, now until June.
Write me if interested.
Cheers
Gio
about publishing data, and fiddling with little
things, and yet it seems there is hardly a system that does any
significant consumption (of any third party data).
Best
Hugh
On 4 Jan 2013, at 21:02, Giovanni Tummarello giovanni.tummare...@deri.org
wrote:
One might just simply stay silent
One might just simply stay silent and move along, but i take a few
seconds to restate the obvious.
It is a fact that Linked data as publish some stuff and they will
come, both new publishers and consumers has failed.
The idea of putting some extra energy would simply be useless per se
BUT it
A good argument ... for using sitemaps·
Yes, those too.
Fundamentally, we need to give discoverability and associated patterns a lot
more focus that has been done in the past. This is such a critical component
for making Linked Data easier to discover and appreciate.
good point re
Generally speaking its yet another big gaping unsolved problem.
Our stab at it one is to do big data Hadoop based summarization and
then use the summary to understand how to query it.
E.g. this unpublished link below exposes the Sindice.com data graph
summary containing a summary of 20-30
Am i really supposed to know if any of the fact below is wrong?
really?
Gio
dbp-owl:PopulatedPlace/area
10.63 (@type = http://dbpedia.org/datatype/squareKilometre)
dbp-owl:abstract
La Chapelle-Saint-Laud is a commune in the Maine-et-Loire department
of western France. (@lang = en)
dbp-owl:area
, schrieb Giovanni Tummarello:
Am i really supposed to know if any of the fact below is wrong?
really?
Its not about factual correctness, but about correct extraction and
representation. If Wikipedia contains false information DBpedia will
too, so we can not change this (at that point). What we want
Sorry all i might be missing a lot of subtleties
are we saying that in the current specs and implementation one can
alter the content of graph B by messing with some triples on a graph A
(one with a blank node?)
Pat i dont get the 'case where subsets of a single large graph are
being
Maybe of interest to some, can be related to data streams,low level
semantic but also to linked data/higher level knowledge
representation.
http://risorseumane.fbk.eu/sites/risorseumane.fbk.eu/files/Call%20TenureTrackICT-BigData.pdf
Gio
: Fri, Sep 28, 2012 17:54
Subject: Expensive links in Linked Data
To: Giovanni Tummarello giovanni.tummare...@deri.org
Cc: Heiko Paulheim paulh...@ke.tu-darmstadt.de, public-lod@w3.org
public-lod@w3.org
Hi,
may I say that the situation you describe is a bit disappointing? The
unaddressed
Short answer is no,
linked data standards have never addressed this and many other even
basic problems(e.g. what if there are too many properties of one kind,
what kind of level of description you're supposed to get (e.g.
recourse on blank nodes?), what is a standard way to find the entry
URI
Sebastian, you might want to use the classic inspector we have at Sindice
http://inspector.sindice.com/inspect?url=http%3A%2F%2Fdbpedia.org%2Fresource%2FLady_Gagacontent=contentType=auto#SIGMA
inspector.sindice.com is used often so you can count on it to be
pretty much ingood working order.
In the past months i have worked a lot on the commercialization of RDF
basedknowledge technologies so i feel like giving a contribution.
We tried to understand what could be of interest to enterprise and
came up with the slogan - or lets say adopted - enterprise linked
data clouds with an
://www.freebase.com/queryeditor
-
Yury Katkov
On Tue, Jul 17, 2012 at 2:30 PM, Giovanni Tummarello
giovanni.tummare...@deri.org wrote:
Thanks for the comments we received.
To answer some of the requests and the will it scale on complex
datasets we have now a sparqled which assists
It might be of interest to some that in Sindice.com we switched from
trying to index all in SPARQL to a mixed approach where all appears on
the frontpage realtime but just selected Websites (rdf,rdfa,
microformats, microdaa etc) + selected LOD datasets appear in a
regularly updated (though not
owned
project. Pls sign up to the google group to express your support.
cheers
Gio
On Sat, Jun 30, 2012 at 6:52 PM, Giovanni Tummarello
giovanni.tummare...@deri.org wrote:
Dear all,
we're happy to release open source today (actually yesterday :) ) a
first version of our data assisted
Dear all,
we're happy to release open source today (actually yesterday :) ) a
first version of our data assisted SPARQL query editor
here is a short blog post which then leads to the homepage and other material
http://www.sindicetech.com/blog/?p=14preview=true
---
Mightwant to try Inspector.sindice.com
might be of SOME use e.g. it will report errors that the w3c wont
report it will support also RDFa of course and other form of data
embedding
might notsupport trig :)
cheers
Gio
On Fri, May 11, 2012 at 11:08 AM, Mark Thompson mark9...@gmail.com wrote:
Hi
HI Peter, all
to add (a probably small element of discussion) to this
i am happy to say that last week we released on the frontpage some
analytics stats which are fresh updated every week.
At the moment they come from 500million+ web URLS. Maybe not much but
pls notice we ONLY retain web urld
Hi lots of work here! Wondering Did you consider evaluating if its worth
learning the quite involved rdf based syntaxes of many of these vs a simple
python jython or whatever script? Obviously you need Turing completeness so
I wonder (and a benchmark should bother to show imo) if its worth the
Tom if you were to do a serious assessment then measuring milliseconds
and redirect hits means looking at a misleading 10% of the problem.
Cognitive loads,economics and perception of benefits are the over the
90% of the question here.
An assessment that could begin describing the issue
* get a
Hugh
here i share my recent experience with a big time (smart) tech
manager of a big time (smart) enterprises we're working on.
* He kept on telling us we're doing liked data, linked data is hot
* I tried to convince him that no.. really liked data is this insane
things that people should use
2012/3/23 Sergio Fernández sergio.fernan...@fundacionctic.org:
Do you really think that base your proposal on the usage on a Powder
annotation is a good idea?
Sorry, but IMHO HttpRange-14 is a good enough agreement.
yup performed brilliantly so far, nothing to say. Industry is flocking
to
me2c
if you can rewrite http://yourserver/page so that it shows as
http://yourpage/resource when page was the result of a redirect that
would indeed finllay resolve the completely unacceptable situations
where users are force to understand (and see in their browser bars)
the distinction.
is the GNU Affero General
Public License, version 3. Sig.ma EE is also available under a
commercial licence. Please contact us to discuss further.
Acknowledgements
Sig.MA EE is built with as part of the LOD2 project. http://lod2.eu
Call: FP7-ICT-2009-5
[1] Giovanni Tummarello, Richard Cyganiak
/ is down..
at least for me.
Juan Sequeda
+1-575-SEQ-UEDA
www.juansequeda.com
On Wed, Aug 24, 2011 at 11:17 AM, Giovanni Tummarello
giovanni.tummare...@deri.org wrote:
Dear all, Sig.ma Enterprise Edition (EE) is now available open source
from http://sig.ma . It can be deployed easily and merge
Seems pretty interesting, clearly out of practical experience !
thanks!
Gio
On Fri, Aug 19, 2011 at 3:56 PM, Leigh Dodds leigh.do...@talis.com wrote:
Hi,
There's a new draft of the Linked Data patterns book available, with
12 new patterns, mainly in the application patterns section.
The
Hi Danny,
i liked sparallax a lot, problem is its hard to maintain. David didnt
upgrade parallax any longer and the intern who did the sparql to MQL
conversion that allows sparallax to operate on sparql is now not
working with us anymore. Hard to say how difficult it would be to
progress on that
Hi out of curiousity
Will you be taking off the diagram those that are NOT online regularly?
Gio
On Tue, Jul 12, 2011 at 7:45 PM, Pablo Mendes pablomen...@gmail.com wrote:
Dear fellow Linked Open Data publishers and consumers,
We are in the process of regenerating the next LOD cloud diagram and
i meant a much simpler and significant thing. Go in CKAN click on the
LOD tag, then start clicking around datasets.
Many dont work, are offline etc. They have been for weeks or months.
Are you checking these and removing them from the new lod diagram or
will the lod diagram just grow regardless
Chris,
i am not interested in specific content of the diagram, but rather i
am interested in understanding what its value of it which depends on
the method you're going to follow in the update. You're answeing this
saying basically there wont be a check for old dead datasets.
I admit never
If you are seeking stats re. what I mean re. intertia, just keep track of
what's happening on the schema.org front re. adoption curve.
here are 100+ datasets
http://sindice.com/search?q=schemanq=fq=class%3Ahttp%3A%2F%2Fschema.org%2F*sortbydate=1facet.field=domaininterface=guru
started
An idea that arose out of a recent discussion with Juergen (in CC): how
about providing a sort of 'bulk data request' facility for your SPARQL
endpoints [1] [2] (as they are, I gather, the more popular ones on the WoD
;)?
Hi wrt to Sindice, it is important that data we index publicly
) channels for data publication
over the web, which serve different goals.
Maybe we need to better articulate the practices and expectations, though...
Cheers,
Antoine
Hi Giovanni,
Le 09/07/2011 23:10, Giovanni Tummarello a écrit :
Hi Nicolas,
Its getting in Sindice indeed -
Yes, I
Hi Nicolas,
Its getting in Sindice indeed - quite politely e.g. 1 every 5 secs-
we'll monitor speed and completeness. iff you think its ok for us to
crawl faster please say so via robot.txt directive or just say so
particular confusion is so destructive. Unlike the dogs-vs-bitches case,
the difference between the document and its topic, the thing, is that one is
ABOUT the other. This is not simply a matter of ignoring some
Could it be exactly the other way around? that documents and things
described in
This year, the Billion Triple Challenge data set consists of 2 billion
triples. The dataset was crawled during May/June 2011 using a random sample
of URIs from the BTC 2010 dataset as seed URIs. Lots of thanks to Andreas
Harth for all his effort put into crawling the web to compile this
Hi Tim ,
documents per se (a la HTTP response 200 response) on the web are less and
less relevant as opposed to the conceptual entities that are represented
by this document and held e.g. as DB records inside CMS, social networks
etc.
e.g. a social network is about people those are the
My sincere congratulations, i had someone overlooked at this level of
detail needed here.
The choices are pragmatic and - in my personal opinion having talked
directly at SemTech with a lot of people involved in this - should
serve the community as good as possible.
will you be posting this as a
my2c
i would seriously advice against using triples with http://schema.rdfs.org .
That would be totally and entirely validating their claim that either
you impose things or fragmentation will distroy everything and that
talking to the community is a waste of time.
For how little this matters
Ireland, Europe
Tel. +353 91 495730
http://linkeddata.deri.ie/
http://sw-app.org/about.html
On 9 Jun 2011, at 09:54, Giovanni Tummarello wrote:
my2c
i would seriously advice against using triples with
http://schema.rdfs.org .
That would be totally and entirely validating their claim
So, can someone clarify, if possible, whether if I publish a page using RDFa
and schema.rdf.org syntax, it will be properly parsed and indexed in any of
those search engines?
that's all they'd have to say not to piss people off but they decided
not to do it.
didnt cost anything. pretty
Hi Frank, my 2c from the Sindice.com point of view.. (as we struggle
to actually make use and make easy for others to use all this)
i wouldn't really worry too much,
just give to the machines what you'd give to humans, that technically
means simply make sure all the pages you display (and that
sindice.com main index has 37,312,159 documents occurrences of foaf:person.
http://sindice.com/search?q=foaf%3Aperson
(a lot of these come from microformats via the any23 library but anyway)
which means there are many more actual persons inside.
Gio
On Wed, Apr 13, 2011 at 10:15 AM, Bernard
, Apr 13, 2011 at 4:48 PM, Giovanni Tummarello
giovanni.tummare...@deri.org wrote:
sindice.com main index has 37,312,159 documents occurrences of foaf:person.
http://sindice.com/search?q=foaf%3Aperson
(a lot of these come from microformats via the any23 library but anyway)
which means
To the best of my knowledge there isnt anything that one could call
modern, updated out there.
something modern and credible would be actual data + social backed
(votes, comments, etc) . . as said in the past we in Sindice we'd be
delighted to provide the data part if anyone wanted to
Boris would you be able to provide a bit of explanation on why would
you want o do that e.g. what evidence are there (nice use cases) were
an rdf export of low level features in the map is of use
thanks!
Gio
On Mon, Jan 17, 2011 at 2:34 AM, Boris Villazón Terrazas
bvilla...@fi.upm.es wrote:
- the rest of the web continue to use 200
Tim
yes but the rest of the web will use 200 also to show what we would
consider 208, e.g.
http://www.rottentomatoes.com/celebrity/antonio_banderas/
see the trilples
Yes Sig.ma heavily checks for properties that are subclass of label
and uses them.
I think sparallax as well.
Gio
On Fri, Nov 12, 2010 at 12:08 PM, Dan Brickley dan...@danbri.org wrote:
Dear all,
The FOAF RDFS/OWL document currently includes the triple
foaf:name rdfs:subPropertyOf
Bravo Harry :-)
let me also add without adding anythng to the header.. *keeping HTTP
completely outside the picture*
http header are for pure optimization issues, almos networking level.
Caching fetching crawling, nothing to do with semantics.
A conjecture: the right howto document is about 2
I might be wrong but I dont like it much . Sindice would index it as 2
documents.
http://iandavis.com/2010/303/toucan
http://iandavis.com/2010/303/toucan.rdf
i *really* would NOT want to different URLs resolving to the same thing
thanks
Giovanni
On Fri, Nov 5, 2010 at 10:43 AM, Ian Davis
How about something that's totally independant from HEADER issues?
think normal people here. absolutely 0 interest to mess with headers
and http responses.. absolutely no business incentive to do it.
as a baseline think someone wanting to annotate with RDFa a hand
crafted, apached served html
Hi Ian
no its not needed see this discussion
http://lists.w3.org/Archives/Public/semantic-web/2007Jul/0086.html
pointing to 203 406 or thers..
..but a number of social community mechanisms will activate if you
bring this up, ranging from russian style you're being antipatriotic
criticizing the
I think it's an orthogonal issue to the one RDFa solves. How should I
use RDFa to respond to requests to http://iandavis.com/id/me which is
a URI that denotes me?
hashless?
mm one could be to return HTML + RDFa describing yourself. add a
triple saying http://iandavis.com/id/me
But again: I agree that crawling the Web of Data and then deriving a dataset
catalog as well as meta-data about the datasets directly from the crawled
data would be clearly preferable and would also scale way better.
Thus: Could please somebody start a crawler and build such a catalog?
As
Hi Mattihas,
sorry for the delay. it is indeed a possible API which we call
longstanding query or notification api . Not yet available , but
we have many requests for it so it wil come.
my advice at the moment would be to do it yourself client side using
say a DB state and fetching the data from
Only solution for you now is to use SPARQL instead of resolving the URI.
Much less traffic and it would actually work
SPARQL doesn't make the problem go away, it just pushes the limits further
out. SPARQL endpoints that see significant traffic have similar restrictions
built in, either on
Thanks Paul, this sort of feedback is indeed tremeoudly useful,
I somehow just wish you had had 1/10th of the replies of the subjects as
literal thread.:-)
Gio
(obviously we're talking business of LOD at large and the true state of it
despite the growing number of lines in the lod cloud diagram.
Jorn you're right.
linked data with plain dereferenciable URIs it plain doesnt work once you
move from the simplest examples. This is for some of the reasons you
mention as well as other others (e.g. how do you really ask what are the
1000 URis most visited (assuming this was in the DB) or the
Apologies for cross posting
-
Dear all
So far semantic web search engines and semantic aggregation services have
been inserting datasets by hand or have been based on random walk like
crawls with no data completeness or freshness guarantees.
After quite some work, we are happy to
Hi there :-) looks very cool.
could you please point us to the specifics of protocol? so we can start
considering integrating in Sindice
Note: we're about to announce (monday?) delta support in Sindice based on
Sitemaps lastmod which seems to be the easiest possible for the HTML+ RDFa
world.
For the interested,
within several new EU projects there are now hiring opportunities
available to work on Sindice current and future services: cloud
computing postdoc/researcher, cloud/semantic/integration developers.
Internships also available with possible ph.d continuation.
Good community
Hi all,
A new version of the Sindice frontend with some interesting improvements.
e.g. a realtime data widget on the homepage, and the new API to
restrict to new day documents (or weekly) etc.
http://sindice.com
Also Facebook support for RDFa is making the web now bubble with new triples.
See
sws.geonames URIs, SPARQL endpoint etc. Bearing in mind that Geonames.org
has no dedicated resources for it, who will care of that in a scalable way?
What is the business model? Good questions. Volunteers, step forward :)
Bernard
Hi Bernard, the need to automatically interlink at large
so hang on tight a bit.. we're working on this, just continue
publishing high quality data with good entity descriptions (as much as
you know about YOUR stuff), and the links will come to you just like
that at some point. I promise :)
WOW ... rings a bell ...and all these things will be
Hi Leigh
i tell you what we're going to be supporting in Sindice very soon and
it would be great if you could add it to the table:
simple existing sitemaps :-). Sitemaps provide the list of URLs to
crawl and for each one either a last updated field or update
frequerncy.
If the website cares to
+1 thanks Nathan for pointing this out, very very relevant.
luckly so far it seems a bit too rooted in MS stack of things (just
looking at it very very superficially) :-)?
Gio
ps: realistically there's the whole microsoft thing to keep in the back
of our minds; they have pretty much a
change or better
reasoning happens or new data etc) + serialization not fully perfomred
automatically would seem an irrealistic
On Wed, Apr 7, 2010 at 12:38 PM, Vasiliy Faronov vfaro...@gmail.com wrote:
Giovanni Tummarello wrote:
In this casematerialization is likely not going to happen much
Wrt this,
i feel like sharing how we address this issue in Sindice and the tools
we provide.
We do materialization at central level following recursively the links
to ontologies e.g. by resolving property names.
This allows data producers to be consideraly more concise in the
markup (e.g. think
Hi Vasily yes, you can use Sindice for that purpose.
either from asking data from the full reasoned cache (ask away ,we
can serve plenty) or from the reasoning API (with a bit of moderation,
it is an intense process although we do have many layers of caching)
a blog post about the details
- general chair Enrico Motta:
http://data.semanticweb.org/person/enrico-motta (see that is general chair
2009)
- a paper from the research track:
http://data.semanticweb.org/conference/iswc/2009/paper/research/311
- a workshop at ISWC2009:
I'd say, if i understand well
that that works only for queries where you need the extra dereferenced
data just additionally e.g. to add a label to your result se
if you need the remote, on the fly reference data to e.g. sort by
price you'd have to fetch all from the remote site ..
Gio
On Sun,
Giovanni Tummarello wrote:
With respect to crawling and scraping or sponging or .. trying to
guess based on partial fragments of structured information i can say
3 thngs
a) No, we're not doing it at the moment, we are only covering those
who chose to put structured semantics. Some book stuff
A) The wrapper's Semantic Sitemap points you at the original Sitemap, and
says how it is doing the wrapping. And because you know how the wrapper is
behaving, you can process the standard Sitemap to get the information you
want about what the wrapping site provides.
Actually, the slicing in
With respect to crawling and scraping or sponging or .. trying to
guess based on partial fragments of structured information i can say
3 thngs
a) No, we're not doing it at the moment, we are only covering those
who chose to put structured semantics. Some book stuff shows up in
Sig.ma .. e.g.
Kind of make me thing.. we could put it virtually back in the same
place as originally on our Sindice cache [1]
i wonder if the operation make sense.. on the one hand a chace is
usually intended for reflecting reality on the other i'd see obvious
practical advantages.
Maybe we could offer an
1 - 100 of 157 matches
Mail list logo