Re: IRC channel?

2008-04-02 Thread Kingsley Idehen


Mark Diggory wrote:


Theres an IRC server at w3.org.  Is there any interest in starting a 
chat room for discussing LoD?


-Mark

~
Mark R. Diggory - DSpace Developer and Systems Manager
MIT Libraries, Systems and Technology Services
Massachusetts Institute of Technology









+1 .

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: imdb as linked open data?

2008-04-03 Thread Kingsley Idehen


Hugh Glaser wrote:

On 03/04/2008 12:41, Kingsley Idehen [EMAIL PROTECTED] wrote:

  

Hugh Glaser wrote:


...
  

Hugh




  

Hugh,

This is an example of many to come, where LOD needs to pitch the value
of Linked Data to Information Publishers :-) I think they will
ultimately publish and host their own RDF Linked Data once the intrinsic
value is clear to them.


And when there is also actual extrinsic value? :-)
  

:-)

But yes, and making it easy for them, possibly by actually doing it for
them, is part of the bootstrap process.
  


Of course :-)

The thing I am trying to work out is exactly how to make the pitch that fits
with their business model, and where their profit line might come from.
This requires a serious understanding of the detailed business model for the
company in question (which is not necessarily a skill the an academic SW
researcher has!).
  


Here is my pitch:
Make yourself more discoverable on the Web via your data .

We also have similar LOD installations for CORDIS (the EU funding agencies'
DB), NSF (a US funding agency), EPSRC (a UK funding agency), and ACM, among
others. We have now engineered them so that they can be moved to the
Information Publisher if desired. Such organisations sometimes have it as
part of their remit to publicise the results, so they should be easier to
deal with, in theory.
  

And they benefit in the same way, in line with goals of their Grants :-)

If anyone has a ready conduit to the appropriate place in such
organisations, we would be delighted to talk with them, showing them what
might be done.
  


Others:  Same here :-)

Kingsley

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com








  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: LOD Community Wiki Space

2008-04-09 Thread Kingsley Idehen


Hausenblas, Michael wrote:

Kingsley,

Great to hear this, thanks! My first Q (sorry, didn't dig deep into it
yet): How can you restrict the access for a page to a community or
individuals (at least for a certain time span), for example to work on a
proposal or coordinate a paper or the like?
  

We can do this via the Admin part of MediaWiki.

I can make you an Administrator if you choose?

Kingsley

Cheers,
Michael
 
--

 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
  
 http://www.joanneum.at/iis/

--
 

  

-Original Message-
From: [EMAIL PROTECTED] 
[mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen

Sent: Tuesday, April 08, 2008 7:01 PM
To: public-lod@w3.org
Cc: public-lod@w3.org
Subject: LOD Community Wiki Space


All,

We know have a MediaWiki instance at:
http://community.linkeddata.org/MediaWiki. 



We can (optionally of course) use this Wiki instance to 
collaborative work on LOD and Linked Data related collateral etc..


The setup is explained at: 
http://community.linkeddata.org/MediaWiki/index.php?VirtuosoWi


ki:About
  

--


Regards,

Kingsley Idehen	  Weblog: 
http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com










  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: ANN: semanticweb.org

2008-05-02 Thread Kingsley Idehen


Dan Brickley wrote:


(resending with right email address for www-rdf-calendar; sorry for 
noise)

Ditto for LOD Community :-)


Dan Brickley wrote:


+cc: RDF calendar list

Denny Vrandečić wrote:

Hello all,

we are proud to announce the relaunch of semanticweb.org

semanticweb.org is a Semantic MediaWiki installation that gathers data
about tools, persons, concepts, events, organizations, and publications
on, around, and about the Semantic Web.

The site provides cool URIs (as certified by the cool URI note authors,
thanks Richard and Leo!) for everything, a SPARQL endpoint, several
exports, and much more. We will see how the data from semanticweb.org
can be reused in your applications.

If you are at the WWW currently, come to tomorrow's presentation of
Kalpana, semanticweb.org and the newest developments on Semantic
MediaWiki, during the Dev Track Sessions from 10:30-12:00. There 
will be
opportunities to ask questions, raise issues, make wishes, 
discussion etc.


Kalpana is a tool for the easy reuse of semantic data inside your 
website.


There will be a more detailed mail next week, if you want more
information before that, pass by tomorrow :)


This is really sweet! Great to see that site come back to live. Wish 
I could've been in China ... if you folks have any trip-reports to 
share, do post them here.


Now, on to the feature requests!

 * http://semanticweb.org/wiki/Events ... can we get iCal feeds?
 * OpenID would be lovely too. Any chance of that?

Re OpenID, I generated a FOAF Group description from the OpenIDs of 
everyone who has edited wiki.foaf-project.org (hmm I should crontab 
it). It is a nice way of building a grassroots list of the OpenIDs of 
people active in the RDF/SWIG and FOAF community. We could also take 
such a list and poke around to find blog RSS feeds etc. to 
auto-generate 'planet' aggregators, lots of possible applications I 
think.


Re Calendar stuff, see http://semanticweb.org/wiki/Talk:Events also 
for requests/suggestions. The Events page is great. And anything we 
can do to move conference CFPs into a more automation-friendly mode 
would be a great help to all our mailboxes I think.


Again, nice work. I look forward to hearing more on what's planned...

cheers,

Dan

--
http://danbri.org/









--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Linked Data talks at XTech in Dublin (6th-9th May)

2008-05-02 Thread Kingsley Idehen


Andrew Walkingshaw wrote:

On Fri, May 2, 2008 at 10:13 AM, Chris Bizer [EMAIL PROTECTED] wrote:
  

 there are plenty of interesting Linked Data talks at XTech this year. For
instance:

 and a complete track on Open Data
 http://2008.xtech.org/public/schedule/topic/23



If you'll excuse the blatant self-promotion, I'm speaking on this kind
of thing too -

http://2008.xtech.org/public/schedule/detail/527
9:45am Friday: Representing, indexing and mining scientific data
using XML and RDF: Golem and CrystalEye  (Andrew Walkingshaw,
University of Cambridge

I hope what I'm going to say will be interesting to Linked Data
people. A large chunk of my presentation's about automated extraction
of triples from a corpus of a few tens of thousands of sets of
experimental data in crystallography, and some of the things you can
do with those triples once you've got them.

Hope to see some of you there!

Best wishes,
Andrew Walkingshaw (http://wwmm.ch.cam.ac.uk/blogs/walkingshaw/)


  

Andrew,

My Linked Data Space 
http://www.openlinksw.com/dataspace/[EMAIL PROTECTED]/weblog/[EMAIL PROTECTED]/1355 
(as in my public Blog Space) has been updated accordingly :-)


Others, you may notice the I am kinda preoccupied with Linked Data 
dog-fooding :-) In this case I've rehashed the mail from Chris in 
palatable Linked Data form via a blog post.


I will be doing the same with the LODW presentations once I have a sense 
of how many are actually live via URIs and how many I need to export to 
the Linked Data Web (via my Briefcase 
http://community.linkeddata.org/dataspace/person/kidehen2 / WebDAV 
space at http://community.linkeddata.org/ods )  from the various mail 
attachments received in response to Chris' initial mailer.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: ESWC 2008 Linked Data Playground

2008-05-29 Thread Kingsley Idehen


David Huynh wrote:


Hi Richard,

If you look at this version

   
http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008-rdf.html


you'll see that the RDF/XML file is linked directly. So, there's 
pretty much zero cost in getting RDF into Exhibit (it took me about 5 
minutes to put that exhibit up). Exhibit automatically routes RDF/XML 
files through Babel for conversion. In the first version, I did that 
conversion and saved the JSON output so that Babel won't get invoked 
every time someone views the exhibit. That's an optimization. Of 
course, Babel isn't perfect in doing the conversion.


Here is an iPhone mockup version for the same exhibit:
   
http://people.csail.mit.edu/dfhuynh/projects/eswc2008/iphone/iphone-exhibit.html 

I only tested it on Firefox and Safari. I think the back button 
functionality doesn't quite work that well, but you get the idea.


David

David,

Even if you don't use RDFa to express what  is about i.e. it's 
foaf:primarytopic, foaf:topic etc..


In the Exhibit pages head/ You can accompany:
link 
href=http://data.semanticweb.org/dumps/conferences/eswc-2008-complete.rdf; 
type=application/rdf+xml rel=exhibit/data /

with
link rel=alternate 
href=http://data.semanticweb.org/dumps/conferences/eswc-2008-complete.rdf; 
type=application/rdf+xml /


I think we need to adopt a multi pronged approach to exposing Linked 
Data (the raw data behind the Web Page):


1. Content Negotiation (where feasible)
2. link rel=.../  (for RDF sniffers/crawlers)
3. RDFa


Re. point 2, I've just taken a random person Abhita Chugh 
http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fsemanticweb.org%2Fwiki%2FAbhita_Chugh 
from http://data.semanticweb.org which exposes the RDF based 
Description of Abhita Chugh 
http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fsemanticweb.org%2Fwiki%2FAbhita_Chugh 
via our RDF Browser without problems (we use all 3 of the methods above 
to seek Linked Data associated with a Web Document). In this case it 
also eliminates the need to translate anything (e.g. routing via Babel)  
since the original data source is actually RDF.


Of course, I could take the exhibit page and slap this in myself, but I 
am hoping you could tweak Exhibit such that it does point 2 and maybe 
point 3 automatically. That would be a major boost re. Exhibit's Linked 
Data credentials :-)


Kingsley


Richard Cyganiak wrote:


David,

On 28 May 2008, at 22:01, David Huynh wrote:

Nice data! I find it useful to look at it this way:

  http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008.html


Very cool! Can you tell us a little bit about the process of getting 
from our RDF to this Exhibit? How much manual/expert work did it 
involve?


(Exhibit is awesome for presenting structured data, and I wonder how 
hard it is in general to get from RDF to Exhibit.)


Cheers,
Richard






Good to see evaluation and interface big enough in the abstract 
word cloud. :-)


David

Chris Bizer wrote:



Hi all,

Paul, Richard, Knud, Tom, Sean, Denny and I have published some 
data describing papers and authors of the 5th European Semantic Web 
Conference (ESWC2008).


The data is interlinked with DBpedia, Revyu and the Semantic Web 
Community Wiki. So if you add a review about a paper to Revyu or if 
you add something to the wiki, your data will mix nicely with the 
data that is already published about the conference.


See http://data.semanticweb.org/conference/eswc/2008/html
for a description of the dataset and its use cases.

The data is currently also being crawled by several Semantic Web 
Search Engines and we hope to be able to add a section about how to 
use the data within these Search Engines before the weekend.


If you have any ideas for further use cases or know about other 
Semantic Web client applications that could be used to navigate, 
visualize or search the data please let me know so that we can add 
them to the webpage.


Have fun playing with the data!

Cheers

Chris


--
Chris Bizer
Freie Universität Berlin
+49 30 838 54057
[EMAIL PROTECTED]
www.bizer.de














--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Contd: ESWC 2008 Linked Data Playground

2008-05-29 Thread Kingsley Idehen


David Huynh wrote:
Exhibit can't do #2 because it only acts on the page at runtime, so 
the author of an exhibit must put that in herself. And that I just did 
for the ESWC 2008 exhibits.





David,

Here are some OpenLink RDF Browser views of your Exhibit generated ESW 
2008 :: Papers 
http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008-rdf.html 
Page, based on your link rel=alternate/ tweak:


1. New Browser Edition View 
http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fpeople.csail.mit.edu%2Fdfhuynh%2Fprojects%2Feswc2008%2Feswc2008-rdf.html;
2. Old Browser Edition View 
http://demo.openlinksw.com/rdfbrowser/?uri=http%3A%2F%2Fpeople.csail.mit.edu%2Fdfhuynh%2Fprojects%2Feswc2008%2Feswc2008-rdf.html


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: More ESWC 2008 Linked Data Playground

2008-05-31 Thread Kingsley Idehen


David Huynh wrote:

Kingsley Idehen wrote:

David,

But should such links be on these pages, too?
Absolutely, in line with the ideal best practice, you should have the 
following routes:


1. Request RDF via Content-type in your HTTP request
2. link rel=-alternate/
3. GRDDL profile in head/ plus link rel=transformation.../
4. eRDF or RDFa

If 1-4 aren't available, Construct a URI that passes the pages 
through an RDFization Service (Babel, Virtuoso Sponger Service, 
Zitgist Services, Triplr, others).

Kingsley

   http://www.eswc2008.org/main_program.html
   http://www.eswc2008.org/program.html
I did a View Source and couldn't find any alternate RDF/XML link. In 
fact, I can't find any RDF/XML on that site. Maybe I wasn't looking 
hard enough.

David,
Then the official ESWC2008 site (eswc2008.org) is not following that 
ideal best practice since it's missing 2, 3, and 4, and as a user with 
just a standard-compliant browser there's little I can do to verify 1. 
Is this diagnose correct? If so, will you please tell the eswc2008 
site admin to get with the program? :-)


Sure!

I am hoping that key principals associated with ESWC 2008  are following 
this thread :-)


We all love to see more open linked data, of course, so it'd be nice 
if *SWC conference sites all lead as examples. In doing so, maybe 
we'll discover the real motivations (or severe lack thereof) for 
publishing SW content, and the challenges of publishing (e.g., tedious 
manual or complex programmatic process to get lat/lng coordinates).


Amen to Dog-fooding!

I hope we are getting closer to the day when the dialog sample below 
becomes the norm:


Technology Vendor or Proponent: I am a vendor and/or proponent of 
Technology X that unveils the virtues of a given paradigm e.g Linked Data


Technology Customer: Do you exploit the virtues of the technology 
yourself? If so, please show me how.


The scenario above is very different from the general practice which 
always omits the vital Dog-fooding aspect :-(


Kingsley


David





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: More ESWC 2008 Linked Data Playground

2008-06-02 Thread Kingsley Idehen


David Huynh wrote:

Thanks, Kingsley. Your responses were crystal clear.

But http://www.eswc2008.org/main_program.html is still devoid of RDF.

Yes :-(

Sean: What's going on with the event's linked data space?


Kingsley


David

Kingsley Idehen wrote:


David,

To this specific point you made:

I believe there need to be a mechanism for rewarding RDF publishing, 
or scolding for forgetting. Do you have that mechanism in-place? 


My response:

1. Reward for RDF publishing  comes down to the benefits or 
opportunity costs associated making a structured data contribution to 
the Web. If you contribute structure you open up the possibility for  
collective participation by others in the Web Community (users and 
developers). If you don't,  then you incur the opportunity costs of 
having to do it all yourself.  Expanding my response is best done by 
reading some of my most recent posts about the emergence of structure 
on the Web in general etc..


2. The mechanism ultimately comes down to degrees to which relevant 
things are discovered in a given space e.g spaces espousing the 
virtues of Linked Data should honor our best practices and radiate 
the values of Linked Data, if they don't, then at the very least, as 
a community, we can flag omissions etc..  I pinged you about a little 
tweak to you exhibit


Now, if my responses are not in line with your question, and I am not 
at ESWC 2008, but absolutely honor the value of discourse driven 
problem resolution, please extend this conversation via some very 
specific suggestions etc..


What would you like to see in place in relation to the questions 
you've posed? Assuming my responses aren't clear enough?








--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Geohash for events

2008-06-09 Thread Kingsley Idehen


Bernard Vatant wrote:


Hi masa

Hi Bernard,

Added support for event page uri:
http://www.kanzaki.com/ns/geo/u07t4qf8j:2008-09-28_2008-10-03;INRIA_IST_2008?uri=http://www.inria.fr/actualites/colloques/2008/ist08/ 

  
Really cool. The URI format looks perfect to me now and exactly what I 
imagined you would do :-) .

The uri part cannot be combined 'hash:datetime;name' part, because
such uri itself adds another hierarchy to the original uri (i.e. / in
event page uri). Hence it should be provided as query string.

I wonder this looks too complicated for practical use ?
  
It is, if users have to concatenate the URI themselves by going to 
geohash, searching the place, copying the geohash id in the service 
namespace, add the time interval in conformant date format, add the 
event URI. Speak about user experience ... for geeks like you and 
me, but ordinary people will never do it that way.


But it you (or someone else) provide a smart bookmarklet (Faviki has 
given me a lot of ideas for that matter) to use in your browser from 
the page of the event 
(http://www.inria.fr/actualites/colloques/2008/ist08/), where the user 
can call geohash via a geocoder, enter the dates using a calendar 
applet, and grab the name from the page title ... et voilà ...
The service would return a page as yours, with the RDF description and 
a permanent URI. And maybe call au passage the geonames service to 
add the neighbouring geonames features, yahoo or google to add 
sponsored links, whatever ...


This would be *practical* ... You could even au passage dump the 
created event in a backend data store, etc.


Say what?


URI are to be felt, experienced, rather than seen. Yes, the extensions, 
plugins, plain old anchor text, and the like are the way to go, the 
Geohase URI scheme are currently implemented is really cool!


I've just opened an example URI in a Browser Session 
http://demo.openlinksw.com/rdfbrowser/?uri=http%3A%2F%2Fwww.kanzaki.com%2Fns%2Fgeo%2Fu07t4qf8j%3A2008-09-28_2008-10-03%3BINRIA_IST_2008%3Furi%3Dhttp%3A%2F%2Fwww.inria.fr%2Factualites%2Fcolloques%2F2008%2Fist08%2F:-) 
Note, that the Map and TimeLine control provide appropriate views.



Kingsley

2008/6/9, Bernard Vatant [EMAIL PROTECTED]:
 

One thing I was wondering was how to encapsulate the URI of the event
itself, something like (completely incorrect syntax, but you get the 
idea

again)



cheers,

  





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Redirection strategy

2008-06-11 Thread Kingsley Idehen


Richard Light wrote:


In message [EMAIL PROTECTED], Sergio 
Fernández [EMAIL PROTECTED] writes


Hi,


This is a typical abstract URI:

http://collections.wordsworth.org.uk/object/GRMDC.C104.9

Is anyone able to diagnose what might be wrong with my page-serving
strategy?


It worked well for me, see Vapour's report [1] over that URI.


That's very encouraging.


What was
the problem that you saw?


Try:

http://tinyurl.com/3w7wrf

The RDF browser seems to load an RDF graph based on the HTML page, 
rather than the RDF.  It can only be the redirection which is failing: 
if you give it the resource URL you get the correct set of triples.


Richard


Richard,

The output below indicates we have something to look into re. our Browser.

curl -I -H Accept: application/rdf+xml 
http://collections.wordsworth.org.uk/object/1993.25

HTTP/1.1 303 See Other
Date: Thu, 12 Jun 2008 02:10:54 GMT
Server: Microsoft-IIS/6.0
X-Powered-By: ASP.NET
Location: http://collections.wordsworth.org.uk/object/data/1993.25
Connection: close
Accept-Ranges: bytes
Vary: accept
Content-Length: 0
Content-Type: application/rdf+xml;qs=1
Set-Cookie: ASPSESSIONIDSCSAQRAR=FAKPBAEADBPFBFAMOPLBPOAM; path=/
Cache-control: private


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: bbc-programmes.dyndns.org

2008-06-20 Thread Kingsley Idehen


Nicholas Humfrey wrote:

hello,

I am trying to get the work we did on:
http://bbc-programmes.dyndns.org/
live on bbc.co.uk.

Does anyone think that there anything that needs changed/fixed before it
does go live?


At the moment we just have RDF/XML views for Brands/Series/Episodes and
Versions. But plan to is to also have RDF views for for the aggregation
pages (tags, genres, formats, services, schedules...) some time in the
future.


It seems to be hard to find a consensus on use of URIs, but here is how
things are the at moment.

HTML Document: http://www.bbc.co.uk/programmes/b00b07kw
RDF Document: http://www.bbc.co.uk/programmes/b00b07kw.rdf
The thing: http://www.bbc.co.uk/programmes/b00b07kw#episode


When asking for RDF here:
http://www.bbc.co.uk/programmes/b00b07kw#episode
you get 303 redirected here:
http://www.bbc.co.uk/programmes/b00b07kw.rdf
  


Description of Episode I assume is:

http://www.bbc.co.uk/programmes/b00b07kw  (an RDF based information resource 
i.e b00b07kw.rdf
without the .rdf extension)

The use of # cuts out the extra hops associated with a 30X :-)


Example: http://kidehen.idehen.net/dataspace/person/kidehen#this

The information resource hosting the description of This Entity (#this) is at:  

http://kidehen.idehen.net/dataspace/person/kidehen 


Hope this helps.

Kingsley




Is that sane, or just it infact be a 302?


nick.


http://www.bbc.co.uk/
This e-mail (and any attachments) is confidential and may contain personal 
views which are not the views of the BBC unless specifically stated.
If you have received it in error, please delete it from your system.
Do not use, copy or disclose the information in any way nor act in reliance on 
it and notify the sender immediately.
Please note that the BBC monitors e-mails sent or received.
Further communication will signify your consent to this.



  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: fans of Linked Open Data and the NFL, NHL, or Major League Baseball?

2008-06-21 Thread Kingsley Idehen


Bob DuCharme wrote:

Kingsley Idehen wrote:
Today we have ESPN and co. offering analysis via the traditional 
one-way TV medium. Tomorrow, I envisage a conversation space 
connected by analytic insights from Joe Public the analyst.  Also, 
what's good for sports applies to Politics, Finance, Soap Operas, and 
other realms.


What I'd really like to do as a next step is to identify a Linked Open 
Data advocate who happens to watch a lot of one of the sports for 
which more data is becoming available, and then encourage that person 
to get in touch with one of the stats-oriented communities within that 
sport to help make these stats available as a SPARQL endpoint. If it 
was soccer, Uche would be an obvious candidate, but judging by 
http://www.sportsstandards.org/oc the NHL, NFL, or MLB seem to be the 
best places to start.


At the next Boston/Cambridge/semweb/LOD meetup, I suggest you keep you 
eye out for Red Sox, Bruins, or Pats-themed clothing...


Bob



Bob,

I am a major Soccer, Football (NFL), Basketball (NBA), and Baseball 
(MLB) (play-offs only) fan.


Uche and I share Soccer fan DNA  amongst other things :-)

I've looked at the site you suggested and it provides a number of great 
data sources for LOD and RDB2RDF projects.


We will certainly have a crack at a Linked Data Space based on what's 
available.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Unobtrusive Manifestation of the Linked Data Web

2008-06-22 Thread Kingsley Idehen


All,

In an attempt to kill many birds with a single stone here is a #swig 
logger session link 
http://chatlogs.planetrdf.com/swig/2008-06-22#T20-23-40 covering the 
thinking behind a new extension re. are realising for Firefox 2.x and 
3.x (of course other browsers to follow).


Gist of the matter:
We are adding: View | Linked Data Sources, to browsers, so that Linked 
Data is revealed to Web Users as the 3rd  of the following options:


1. View Rendered Page (what you get by default)
2. View Page Source (how you look at the markup behind the page)
3. View Linked Data Sources (how you look at the raw data behind the 
page; the data the page puts in context in rendered form)


Of course, the same applies to the Browser Context Menus, meaning: 
Linked Data Sources occurs as an option (for completeness this will be 
changed to: View Linked Data Sources).


Our extension uses an HTTP Proxy (Sponger) which is associated with 
RDFizers (think of these as RDF analogs of ODBC/JDBC Drivers, but for 
Web Data). Thus, you can increase or decrease the quality of the linked 
data graph via the Cartridges/Drivers that you plug into your data spaces.


Also note, the recent enhancement to Semantic Radar now makes it 
possible to incorporate Linked Data Browsers/Viewers from a myriad of 
providers (OpenLink, Zitgist, Marbles, DISCO, Tabulator etc.), which is 
great, but you will note that the new plugin doesn't mandate any 
discovery of RDF. It simply takes a URL and then RDfizes it whenever you 
take the: View | Linked Data Sources menu or context menu route.


The initial cut of the extension is at:
http://myopenlink.net:8890/~kidehen/Public/rdfb.xpi

Of course, there are a few updates on the way later this week(mostly 
aesthetic bar the non functioning Search feature). That said, you can 
get to the essence of what I am talking about via the initial release.


Related:

  1. My Linked Data Planet Keynote
 
http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2.html
 (thanks to RDFa and the Bibliographic Ontology it's now possible
 to model Presentations (Slideshows) )
  2. Other Slidy Presentations from OpenLink
 http://virtuoso.openlinksw.com/Whitepapers/index.htm enhanced
 with RDFa annotations.


Final note: since we already have a Cartridge/Driver for Slidy based 
data sources, you can also view the data behind any Slidy presentation. 
Of course, the context fidelity of the graph would be low (i.e. low 
linked data resolution) without RDFa or some other Linked Data exposure 
mechanism, but if you have a meta cartridge (e.g. the kind that Zitgist 
delivers via it's exploitation of UMBEL aided Named Entity Extraction 
and Disambiguation), you don't even require RDFa en route to context 
fidelity (HIgh Def. Linked Data), you just slot that cartridge into your 
data space.


BTW - http://community.linkeddata.org/ods remains open to community 
members seeking both a data space and an easy place to get an Entity URI :-)


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Contd: Unobtrusive Manifestation of the Linked Data Web

2008-06-22 Thread Kingsley Idehen


All,

Quick note re. the Firefox Linked Data Sources extension ( rdfb.xpi) :

1. Download to your local filesystem
2. Use File | Open from Firefox to load  (this isn't a properly signed 
XPI at the current time)



Kingsley


All,

In an attempt to kill many birds with a single stone here is a #swig 
logger session link 
http://chatlogs.planetrdf.com/swig/2008-06-22#T20-23-40 covering the 
thinking behind a new extension re. are realising for Firefox 2.x and 
3.x (of course other browsers to follow).


Gist of the matter:
We are adding: View | Linked Data Sources, to browsers, so that Linked 
Data is revealed to Web Users as the 3rd  of the following options:


1. View Rendered Page (what you get by default)
2. View Page Source (how you look at the markup behind the page)
3. View Linked Data Sources (how you look at the raw data behind the 
page; the data the page puts in context in rendered form)


Of course, the same applies to the Browser Context Menus, meaning: 
Linked Data Sources occurs as an option (for completeness this will 
be changed to: View Linked Data Sources).


Our extension uses an HTTP Proxy (Sponger) which is associated with 
RDFizers (think of these as RDF analogs of ODBC/JDBC Drivers, but for 
Web Data). Thus, you can increase or decrease the quality of the 
linked data graph via the Cartridges/Drivers that you plug into your 
data spaces.


Also note, the recent enhancement to Semantic Radar now makes it 
possible to incorporate Linked Data Browsers/Viewers from a myriad of 
providers (OpenLink, Zitgist, Marbles, DISCO, Tabulator etc.), which 
is great, but you will note that the new plugin doesn't mandate any 
discovery of RDF. It simply takes a URL and then RDfizes it whenever 
you take the: View | Linked Data Sources menu or context menu route.


The initial cut of the extension is at:
http://myopenlink.net:8890/~kidehen/Public/rdfb.xpi

Of course, there are a few updates on the way later this week(mostly 
aesthetic bar the non functioning Search feature). That said, you 
can get to the essence of what I am talking about via the initial 
release.


Related:

  1. My Linked Data Planet Keynote
 
http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2.html 


 (thanks to RDFa and the Bibliographic Ontology it's now possible
 to model Presentations (Slideshows) )
  2. Other Slidy Presentations from OpenLink
 http://virtuoso.openlinksw.com/Whitepapers/index.htm enhanced
 with RDFa annotations.


Final note: since we already have a Cartridge/Driver for Slidy based 
data sources, you can also view the data behind any Slidy 
presentation. Of course, the context fidelity of the graph would be 
low (i.e. low linked data resolution) without RDFa or some other 
Linked Data exposure mechanism, but if you have a meta cartridge (e.g. 
the kind that Zitgist delivers via it's exploitation of UMBEL aided 
Named Entity Extraction and Disambiguation), you don't even require 
RDFa en route to context fidelity (HIgh Def. Linked Data), you just 
slot that cartridge into your data space.


BTW - http://community.linkeddata.org/ods remains open to community 
members seeking both a data space and an easy place to get an Entity 
URI :-)





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Linked Data re Non-Profits and NGO's. Have data, need vocabulary.

2008-06-27 Thread Kingsley Idehen


Chris Bizer wrote:

Hi Bob and Kingsley,
 
a while ago there was a RDF version of Wikicompany online at 
http://dbpedia.openlinksw.com/wikicompany/resource/Wikicompany
 
Maybe it would also be an idea to reuse terms from the vocabulary of 
this source.
 
Kingsley: The URI abouve currently gives a 500 return code. Do you 
know what happend to the site?
 
As an alternative you could also think about reusing the terms that 
are currently used within DBpedia to describe organizations.

Chris,

Wikicompany is certainly the place to start.

The URI should be live again.

Kingsley


 
Cheers
 
Chris
 
--

Chris Bizer
Freie Universität Berlin
+49 30 838 54057
[EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
www.bizer.de http://www.bizer.de

- Original Message -
*From:* Bob Wyman mailto:[EMAIL PROTECTED]
*To:* public-lod@w3.org mailto:public-lod@w3.org
*Sent:* Friday, June 27, 2008 1:19 AM
*Subject:* Linked Data re Non-Profits and NGO's. Have data, need
vocabulary.

I would like to make available as Linked Data several databases
describing several million non-profits, NGO's and foundations. The
data includes things like name of organization, address, budget,
source of funds, major programs, key personnel, relationships to
other organizations, area of expertise, etc.

What I don't have is an RDF vocabulary with which to describe
these things. While I could define one myself, I would like to
base my work on existing standards, or common practice, however,
seemingly endless digging through the web indicates that there
aren't any obvious standards for describing even basic things
like address in RDF. Perhaps, I'm looking in the wrong places...

Ideally, I would find some well formed vocabulary for a Business
Description that I could use or adapt. I would appreciate it if
anyone could give me pointers to either such a well worked
vocabulary or at least to smaller vocabularies for things like
address that I could use in composing a vocabulary with which to
publish this data. Can you help?

bob wyman




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)

2008-07-14 Thread Kingsley Idehen


Tom Heath wrote:

-Original Message-
From: [EMAIL PROTECTED] 
[mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen

Sent: 12 July 2008 21:43
To: [EMAIL PROTECTED]
Cc: public-lod@w3.org; [EMAIL PROTECTED]
Subject: Re: Ordnance Survey data as Linked Data (RE: How do 
you deprecate URIs? Re: OWL-DL and linked data)



Kingsley Idehen wrote:

I also forgot to mention obvous use of RDFa in the HTML doc 
which broadens the range of rdf aware user agents tha 
commence RDF discovery from HTML



Question: is it worth creating a duplicate RDF graph by using RDFa in
HTML documents, when there is also RDF/XML available just one link
rel=.../ away, and at a distinct URI? Doesn't this RDFa + RDF/XML
pattern complicate the RDF-consumption picture in general if we assume
agents will want to do something with data aggregated from a number of
sources/locations, i.e. doesn't it increase the cost of removing
duplicate statements by creating more in the first place? Does it not
also complicate the picture of making provenance statements using named
graphs, if the subject of the triple could be both an HTML document and
an RDF graph?

Dunno the answers to these questions, but interested to hear what people
think.

Tom.

  

Tom,

I believe we should spread the net as wide as possible re. RDF aware 
user agents since we cannot assume which methods they will be using to 
discover RDF.  On our part (re. our RDF aware user agents), we use all 
the methods (Content Negotiation, link /, GRDDL, RDFa, and POSH) when 
sniffing, which also implies that we

take on the burden of normalization.

On the part of publishers, I encourage the use of at least one of the 
RDF exposure methods mentioned above, when known associations between 
HTML and RDF representations exist.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)

2008-07-14 Thread Kingsley Idehen


Hugh Glaser wrote:

Thanks Tom.
Er, yes.
I was puzzled by the suggestion that I might duplicate the RDF in the page that 
did a simple html rendering of the underlying RDF I was trying to publish.
I would have thought that this is actually a Bad Thing, rather than a Good 
Thing.

And if we are talking about an RDF browser (as our pages are, albeit with a 
clean URI that doesn't have the browser URI in it), getting it to include the 
RDF as RDFa or whatever is even stranger; after all
http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fos.rkbexplorer.com%2Fdescription%2Fosr70017765
doesn't include the substantive RDF as RDFa, (or have a link rel to 
http://os.rkbexplorer.com/data/osr70017765 for that matter) which would 
be the equivalent.
  

Hugh,

I knew this was coming.

Please practice what I say and not the current state of our user agents :-)

I prefer to make suggestions that go beyond what we've implemented. Yes, 
of course, I advocate dog-fooding but I also have to deal with the 
realities of development and product release cycles etc..


Our agents will be fixed in line with my suggestions, for sure. 

My key point is this: we are a  community of knowledgeable folks who 
need to take on the the burden of unobtrusive injection of  RDF into the 
Web. We cannot expect this to happen outside the community at this 
stage. Using HTML as the vehicle for RDF exposure is a big deal and 
offers immense value. RDFa is a major contribution to the whole RDF 
exposure puzzle, and this is one area where it's value is crystal clear  
(imho).


Kingsley

On 14/07/2008 09:55, Tom Heath [EMAIL PROTECTED] wrote:



  

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen
Sent: 12 July 2008 21:43
To: [EMAIL PROTECTED]
Cc: public-lod@w3.org; [EMAIL PROTECTED]
Subject: Re: Ordnance Survey data as Linked Data (RE: How do
you deprecate URIs? Re: OWL-DL and linked data)


Kingsley Idehen wrote:

I also forgot to mention obvous use of RDFa in the HTML doc
which broadens the range of rdf aware user agents tha
commence RDF discovery from HTML



Question: is it worth creating a duplicate RDF graph by using RDFa in
HTML documents, when there is also RDF/XML available just one link
rel=.../ away, and at a distinct URI? Doesn't this RDFa + RDF/XML
pattern complicate the RDF-consumption picture in general if we assume
agents will want to do something with data aggregated from a number of
sources/locations, i.e. doesn't it increase the cost of removing
duplicate statements by creating more in the first place? Does it not
also complicate the picture of making provenance statements using named
graphs, if the subject of the triple could be both an HTML document and
an RDF graph?

Dunno the answers to these questions, but interested to hear what people
think.

Tom.

--
Tom Heath
Researcher
Platform Team
Talis Information Ltd
T: 0870 400 5000
W: http://www.talis.com/platform




  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)

2008-07-14 Thread Kingsley Idehen


Hugh Glaser wrote:

On 14/07/2008 10:42, Mark Birbeck [EMAIL PROTECTED] wrote:
[snip]
  


And if we are talking about an RDF browser (as our pages are, albeit with a
clean URI
that doesn't have the browser URI in it), getting it to include the RDF as
RDFa or whatever
is even stranger; after all
http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fos.rkbexplorer
.com%2Fdescription%2Fosr70017765
doesn't include the substantive RDF as RDFa, (or have a link rel to
http://os.rkbexplorer.com/data/osr70017765 for that matter) which
would be the
equivalent.
  

I can't comment on that example, but ultimately there is no need for a
URL for an HTML+RDFa page to be any different to a normal one.

(Although I might have missed your point, here)


Not sure.
I think it relates to the question Tom is asking.

Another way of putting it is that we don't expect Tabulator to include RDFa
of the RDF we are currently browsing (I think - or maybe we should?).
  

Hugh,


So I think it comes down the purpose of the exercise, and there is no one
size fits all.

If I have some web pages, then being able to simply embed RDF (by whatever
means), or link to associated RDF resources, is really great. Good for
convenience, maintenance.

On the other hand, if I am trying to simply publish RDF, for example out of
a DB, then things are a bit different. To help people who want to build
agents that use it I might expect them to be able to visualise what I am
publishing by using Tabulator or other tools to browse it generically. Or
equivalently I might provide some human readable pages of the RDF to make my
data more easily accessible.
Because it is a bespoke browser, then I need to add rel link to the RDF URI.
And then someone suggests I should add RDFa or something more to these
pages.
Before I know it, I have invested a lot of time providing (and maintaining)
neatly formatted and RDF-friendly web pages to publish my DB, when all I
wanted to do was join the Semantic Web by publishing my RDF.

In fact, one of the options when Kingsley asked for the rel link was to
simply remove all the html description pages, getting back to the core of
what I want to do, which is publish RDF as Linked Data.
  
Yes, and the little link rel=alternate /  addition has extended the 
scope of your Linked Data deployment to the broader Web :-) And if you 
choose, you can broaden even further will some RDFa sprinkled in.  Note, 
this is simply an option.




Kingsley

Best
Hugh



  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Linked Movie DataBase

2008-08-01 Thread Kingsley Idehen


Richard Light wrote:


In message [EMAIL PROTECTED], Jens Lehmann 
[EMAIL PROTECTED] writes


How do I get it to give me XML RDF?  I only seem to be able to get 
these

text triple things, which I don't consider to be a machine-processible
format ;-)


I get XML when requested. Try:

curl -H accept: application/rdf+xml
http://data.linkedmdb.org/resource/film/2014 -L


Jens,

Yes, thanks - so do I.

I was trying to get XML RDF by using Firefox with the Tabulator 
extension, and just expected it to kick in when I went to the data 
URL.  Instead I found myself looking at these text triples.


Is there a variant on the URL itself which will return XML RDF?  My 
concern is that XSLT processors should be able to access these linked 
data resources. XSLT [1.0] only groks XML documents.  XSLT just has 
the document() function in its armoury, and the only information you 
can give it is a URL: hence the question.


(I took this concern to the XSLT list, and the most helpful suggestion 
I got there was to set up a proxy server which takes URLs, adds an 
accept header to them, and passes them on.  So that's my fallback 
strategy if I can't put something into the URL to achieve the desired 
result.)


Richard


Richard,

Please try the Open Data Explorer (esp. Firefox Extension variant [1] ), 
it works fine with the Linked Movie Database as per some of my twitter 
micro posts from yesterday [2] . You will also see examples re. other 
exciting Linked Data additions such as data from the BBC and Southampton 
Pubs.  I even drop a simple mashup in one the links to boot.


Links:
1. https://addons.mozilla.org/en-US/firefox/addon/8062
2. http://twitter.com/kidehen



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Visualizing LOD Linkage

2008-08-06 Thread Kingsley Idehen


Peter Ansell wrote:

- Yves Raimond [EMAIL PROTECTED] wrote:

  

Hello!



It depends on whether you know that the external references are
  

distinct just based on the URI string. If someone links out to
multiple formats using external resource links then they would have to
be counted as multiple links as you have no way of knowing that they
are different, except in the case where you reserve these types of
links as RDF literals.

I am not sure if I interpret it correctly - do you mean that you

could
link to two URIs which are in fact sameAs in the target dataset?
Indeed, in that case, the measure would be slightly higher than what
it should be. However, I would think that it is rarely (if not never)
the case.



  

Peter,

I personally don't put sameAs on URI's which relate to the same thing but are 
really just different representations, ie, the HTML version doesn't get sameAs 
the RDF version.
  


I really don't believe anyone in this community advocates using 
owl:sameAs between representations. We use owl:sameAs between Entity 
URIs while representations are negotiated (content negotiation), 
discovered via link rel=../, or RDFized etc..


If people knowingly mapped owl:sameAs between representations that 
weren't identical, then of course this would be flawed if the 
representations weren't identical. But this isn't what's happening in 
the LOD space in general,  or it's flagship effort: DBpedia.


http://dbpedia.org/resource/Berlin is not the representation of the 
entity Berlin, it's a pointer (Identifier) used  by the deploying 
platform to transmit the description of said entity using a 
representations desired by the requester/consumer/agent .  This entire 
mechanism isn't new to computing, it's how all our programs work at the 
lowest levels i.e., we interact with data by reference using pointers [2].


This matter is heart and soul of linked data on the Web or across any 
other computing medium that manipulates data.


Links:

1. http://en.wikipedia.org/wiki/Dereferencable_Uniform_Resource_Identifiers
2. http://cslibrary.stanford.edu/104/  (which clearly needs  Linked Data 
Web variant)


Kingsley

 
  
[SNIP]

(link appears to be broken)



Springer Link DOI system must be broken. Try the following

http://www.springerlink.com/content/w611j82w7v4672r3/fulltext.pdf

The following image link is also quite interesting:

http://bio2rdf.wiki.sourceforge.net/space/showimage/bio2rdfmap_blanc.png

Cheers,

Peter

  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Fwd: [backstage] Muddy Boots + BBC Music Beta

2008-08-06 Thread Kingsley Idehen


Chris Sizemore wrote:


as i say, the Muddy Boots folks are very LOD-centric and 
dbpedia-sympathetic, so we should just mention these constructive 
criticisms to them and i'm sure they'll be taken seriously.


also to note is that their work is the result of a BBC commission that 
i was involved with, so i expect the BBC to make the links between 
News stories and dbpedia/musicbrainz available itself, eventually. 
also, the Muddyboots entity extraction source code will be available 
as open source, i do believe.


as rob mentions in his original email, this is proof-of-concept stuff, 
and linked data wasn't the main use case (though linked data should 
drop out of it naturally...)



Chris,

Sure re. these folks, but there are many others that consuming these 
URIs (as we can see from our logs) .


The great thing about URI based Attribution is the fact that you can 
walk the graphs exposed via the HTTP logs, and check adherence to 
attribution methods.


Lossy attribution value chains remain the biggest hurdle to Linked Data 
deployment.


If data publishers understand the intrinsic power of Linked Data (i.e 
brand imprint via URIs) combined licensing that requires URI based 
Attribution, they will be more inclined to publish web documents that 
expose Linked Data; especially as URI endowed entities distilled from 
their Web ultimately extend/expand/explode the range of their value 
propositions.


If Attribution by URI isn't unwieldy, then we have a business model 
unraveling game-changer that extends even beyond the early adapter 
echelons of LOD  :-)


Kingsley




best--

--cs


-Original Message-
From: [EMAIL PROTECTED] on behalf of Kingsley Idehen
Sent: Wed 8/6/2008 5:19 PM
To: Yves Raimond
Cc: public-lod@w3.org
Subject: Re: Fwd: [backstage] Muddy Boots + BBC Music Beta


Yves Raimond wrote:
 Hello!

 I thought this would be interesting for this list, as an example of a
 service using several LOD sources.
 However, it is a bit sad they don't expose the data they produce as
 linked data themselves - perhaps we should have a GPL-like license for
 LOD datasets if you derive data from this linked data, it must be
 available as linked data :-D (just jocking).

 Cheers!
 y


 -- Forwarded message --
 From: robl [EMAIL PROTECTED]
 Date: Wed, Aug 6, 2008 at 4:15 PM
 Subject: [backstage] Muddy Boots + BBC Music Beta
 To: [EMAIL PROTECTED]


 Hi,

 Over at Rattle [1] we got quite excited by the release of the new
 music beta site, in fact we created our own prototype that links our
 latest iteration of the Muddy Boots system [2] with the beta music
 site.

 We haven't announced the latest version of Muddy Boots yet, but in
 essence it's main aim is to 'unambiguously identify the main actors in
 a BBC news story', in doing this it uses DBpedia URI's to identify the
 entities involved. At the moment the system knows about 'people' and
 companies' however we've just added experimental support for 'bands'
 (!).

 As DBpedia knows about Musicbrainz guid's and Muddy Boots knows which
 BBC news stories relate to which DBpedia entry for a person|band we
 can add metadata about the related news stories about an artist ('Seen
 in these news stories' at the bottom of the page) :

 Coldplay : 
http://muddy.rattleresearch.com/muddy2/musicentities/cc197bad-dc9c-440d-a5b5-d52ba2e14234


 Kaiser Chiefs :
 
http://muddy.rattleresearch.com/muddy2/musicentities/90218af4-4d58-4821-8d41-2ee295ebbe21


 and of course not forgetting the famous spoken word artist (!) :

 George W Bush :
 
http://muddy.rattleresearch.com/muddy2/musicentities/06564917-bdd2-4fb6-bcdc-be9e0c04f7ac


 We'll be mentioning more about Muddy Boots in the future, but you can
 see we're starting to add semantic markup to BBC News stories and
 creating links between open data sources and BBC News (try clicking a
 news story and you'll see it has the 'actors' in the news story marked
 up with Microformats). It's still at the prototype stage at the moment
 and we're about to enter a formal validation and testing phase to
 measure the systems accuracy - but we thought we'd produce this
 prototype to demonstrate the kinds of things we can start to achieve
 when open data (or web-scale) identifiers are used to identify
 content.

 You can see all the music entities the system knows about by viewing :
 http://muddy.rattleresearch.com/muddy2/musicentities/

 The support for identifying bands is definitely considered
 'experimental' at the moment, so you might see the occasional 'blip'
 with related stories as we look at how to classify 'bands' in stories
 more accurately. We're just started indexing BBC news stories in
 anger, so expect to see more related data appear over the next fews
 days and weeks.

 We'd love to hear any comments you have about it :)

 Thanks,

 Rob

 [1] http://www.rattleresearch.com
 [2] http://muddyboots.rattleresearch.com/semantic-web-project/
 -
 Sent via the backstage.bbc.co.uk discussion group.  To unsubscribe

Re: freebase parallax: user interface for browsing graphs of data

2008-08-13 Thread Kingsley Idehen


Richard H. McCullough wrote:


Very nice!!!

How is your knowledge base structured?  What language?

Can I download your program to my computer?

Dick McCullough
Ayn Rand do speak od mKR done;
mKE do enhance od Real Intelligence done;
knowledge := man do identify od existent done;
knowledge haspart proposition list;
http://mKRmKE.org/

David,

For purpose of clarity and broader discourse (I know we've been over 
this in private), what is the Linked Data and/or Semantic Web oriented 
value of this very cool visualization?


Put differently, it would be nice if I could beam a query down the data 
graph exposed by you very nice visualization rather than being confined 
to the options presented by your application.


At the very least, what's the harm in exposing the Freebase URLs in 
these Web Pages? If you do that, at the very least, other user agents 
can do stuff with the graphs (Linked Data) that you are visualizing.  
The cost of this little tweak is extremely low and the upside extremely 
high.


Cool stuff for sure, but I would like Parallax to be a nice Linked Data 
Web contribution also :-)  Freebase (basic) is part of the Web, but 
Parallax is sort of confining me to a Freebase enclave by not exposing 
URLs (where I currently see: javascript:{}).



My fundamental argument remains this:

Visualizations are good, but graph visualizations are not the sole keys 
to the treasures that reside within a Graph. Effective traversal (e.g. 
query beaming SPARQL or MQL) is also part of the puzzle, and it would be 
nice if we could always offer the visualization and the 
graph-query-beam-beachhead as part of a single Web information resource 
deliverable. There are always reasons why one or more humans (due 
inherent cognitive prowess) would seek to view the same data 
differently, no matter how compelling the initial visualization, due to 
the fact that data visualizations are inherently subjective projections.


Links:

1. http://tinyurl.com/5of2qu - Abraham Lincoln as Linked Data from 
Freebase  ( I get no triples with Parallax pages since the Freebase Data 
Sources aren't exposed)


Note: The Freebase RDFization Cartridge (Wrapper, Scrapper etc.) will be 
better i.e., right now there are too many literal property values that 
should be URIs (this fix is coming). Ditto proper lookup driven meshing 
with DBpedia.



Kingsley


- Original Message - From: David Huynh [EMAIL PROTECTED]
To: public-lod@w3.org; [EMAIL PROTECTED]
Sent: Tuesday, August 12, 2008 3:11 PM
Subject: freebase parallax: user interface for browsing graphs of data




Hi all,

I've been exploring some user interface ideas for browsing graphs (of 
data in Freebase) in a user-friendly way, which I think might be 
applicable to some of the data that you have. The screencast should 
explain:


   http://mqlx.com/~david/parallax/

Please let me know what you think!

Thank you.

David












--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: freebase parallax: user interface for browsing graphs of data

2008-08-14 Thread Kingsley Idehen


David Huynh wrote:

Kingsley Idehen wrote:

David,

For purpose of clarity and broader discourse (I know we've been over 
this in private), what is the Linked Data and/or Semantic Web 
oriented value of this very cool visualization?

Hi Kingsley,


David,
I apologize for the confusion. I actually didn't say semantic web or 
web of data or linked open data in my first message.
I know, but linkage is inferred by virtue of the forum posts :-)  I 
think Sandro Hawke had a similar response re. congruence.
I simply thought that the graph-based browsing paradigm of Parallax 
might be useful on RDF, which is graph-like. Dbpedia, for example, is 
a huge graph of billions of triples, and as far as I'm aware, it's 
hard to explore Dbpedia.

I have not doubt about its virtues on the browsing side, none whatsoever.


And there has been very similar UI research on browsing graphs, such 
as Georgi's Humboldt, which I missed out as I wasn't able to attend 
WWW 2008 :-( Well, at least I had a chance to discuss with him on the 
Simile mailing list before.


But if you would still like to understand the relevance to Linked Data 
/ SW, then may I point you to third parties who have tried to 
ask/answer that question:


   http://blogs.zdnet.com/collaboration/?p=131

Excerpt:


W3C director Sir Tim Berners-Lee’s long term vision of the Web as a 
universal medium for data, information, and knowledge exchange would 
make it possible for the web to understand and satisfy the requests of 
people and machines.



Keyword: Exchange.
Key Mechanism: URIs or URLs.

All I ask of you is the exposure of the Freebase URLs that are already 
in sight re. the current / basic Freebase pages.  javascript:{} is 
neither URL nor URI.



The ‘people operated’ side of things is the world we live in today. As 
David Huynh’s video discusses, we busy ourselves manually searching by 
keyword in multiple locations and then compile our results.



Yes, we may be less busy as a result of nice UI (visualization or 
interaction) but the ultimate goal is stimulation of our inherent 
cognitive prowess. Thus,  also offer users of these UIs  an optional 
route to the raw data. On the Web that's simply about URIs or URLs. A 
stimulated human is an inquisitive human, and an inquisitive human will 
ultimately seek to explore dimensions associated with observation, 
perception, and comprehension.



  (the talkback comments are also valuable)
Can't comment on zdnet as they still require completion of a full 
subscription interview first, no OpenID or the like in sight :-(




   
http://www.readwriteweb.com/archives/freebase_parallax_taunts_us_wi.php
No comment, I think that article speaks for itself. From my vantage 
point, readwriteweb is still grappling with the fundamentals of what 
comes after REST APIs.




   
http://blog.futurefacts.net/2008/08/14/future-web-freebase-%E2%80%93-killing-google-with-semantic-precision/ 

That article is another variant of the readwrite article re. overall 
comprehension.


None of them answered the question re. Linked or Semantic Web relevance, 
because none of them seem to understand the essence of either. For 
instance, they don't see how Networking is the constant while points 
of reference (network entry points) vary over time; we started with 
Routers (eons ago) and now we are dealing with Entities across Data 
Spaces [1][2].


Those are their opinions alone. I include them here to provide various 
perspectives.


Here is my basic opinion:

I don't have an issue with the UI, it's nice, cool, and innovative. I 
just want it to be part of the Web. Being part of the Web is about 
accepting that Linkage is sacrosanct. Thus, no silos, no matter how 
pretty, cool, or innovative.


Your visualization is how I would lead a user to a beachead point in the 
underlying graph so that they could beam a SPARQL + Full Text pattern 
query down the graph without them ever writing or seeing a line of SPARQL.


Question: Is there any fundamental reason why you cannot expose URIs or 
URLs where you have javascript:{}? Would this break your work in 
anyway?  If you expose the URIs, I would then be able to demonstrate 
what I mean using your nice UI, how about that?


Links:

1.  
http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2_TimBL_v3.html#(15) 
-- Start
2. 
http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2_TimBL_v3.html#(22) 
-- Where we are headed


Note:
The presentation above is a remix of presentations by  TimBL and I from 
the recent Linked Data Planet conference. You've forced my hand re. 
publication as this remix is also about demonstrating RDFa virtues 
amongst other things re. the Linked Data Web  :-)




Best,

David





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http

Re: freebase parallax: user interface for browsing graphs of data

2008-08-14 Thread Kingsley Idehen


Georgi Kobilarov wrote:

Hi Kingsley,

  

Is this an apropos or an insight addition?

My entire position about Linked Data solutions is centered on User
Interaction.  Note, I said: Visualizations are good, but graph
visualizations are not the sole keys to the treasures that reside
within
a Graph.



True.

Well, my point was that I got the impression that you've characterized
David's interface as a graph visualizer. Which is, imho, not correct.
  
My only issue with Davids work is: javascript:{} where there should be 
a URI or URL. That's it.



And I'd like to highlight the distinction between visualizations and
interaction models. The core of every interface is the interaction
model. On top of that, there might be specific visualizations. 


The problem is that most graph UIs I know are based on the one resource
at a time interaction model. It's the interaction model of the current
web, where a resource is a web page, and users interact by looking at
one page (reading it), and browsing from one page to another. That
interaction model underlies all current linked data browsers (Tabulator,
Disco, etc.). Some UIs use a graph visualization on top, but the
interaction model remains the same. (As a side note, I think that
visualizing a data as a graph is useless. Why using a 2-dimensional
layout where both dimensions are undefined?) 


But the real magic happens when you change to interaction model, and
users can interact with multiple resources at a time. David has shown
some excellent examples. Exhibit provided a faceted filtering
interaction for graph data. And Parallax now demonstrates a solution to
graph browsing.

  

That is part of the magic. There are no panaceas.

Another part of the magic is the freedom to flip modes i.e., access to 
data behind the pretty pages (or other modes of UI).


  

Give users the  option to use their own cognitive skills to interact
with the data (e.g. pivot, project, and traverse based on their own
specific needs).



Give users as few options as possible, and design them in a way, that
users do not have to use their cognitive skills.
  

Please re-read your comments.

In my world view, I don't assume I am smarter than your hypothetical 
Users.


In my world view, I seek to stimulate and arouse the cognitive skills of 
the users (those who encounter my work).


In my world view, I don't assume I have all the answers.

In my world view, I believe that harnessing collective intelligence is 
the ultimate gift of the Web.


Cognition is what makes our world tick. It's what guarantees innovation 
in a timeless continuum.  People have to think outside the box or we 
stagnate and then regress.



  

Programmers cannot acquire the domain expertise
of an alien domain.



They should. Or at least have an interaction designer supporting them.
  
An excellent programmer and an excellent interaction designer cannot 
surmount stimulating cognitive power. What you describe is a futile 
quest (imho).  Ironically, you left out systems designers, architects, 
and data analysts amongst others. Even the sum of all of the missing 
parts still won't get you close.



Kingsley




Best,
Georgi

--
Georgi Kobilarov
Freie Universität Berlin
www.georgikobilarov.com

  

-Original Message-
From: Kingsley Idehen [mailto:[EMAIL PROTECTED]
Sent: Thursday, August 14, 2008 2:25 PM
To: Georgi Kobilarov
Cc: public-lod@w3.org; [EMAIL PROTECTED]
Subject: Re: freebase parallax: user interface for browsing graphs of
data

Georgi Kobilarov wrote:


Kingsley,


  

Visualizations are good, but graph visualizations are not the sole



keys

  

to the treasures that reside within a Graph.



It's not about graph visualization. It's about user interaction for
graph-based data.

  

Georgi,

Is this an apropos or an insight addition?

My entire position about Linked Data solutions is centered on User
Interaction.  Note, I said: Visualizations are good, but graph
visualizations are not the sole keys to the treasures that reside
within
a Graph.

I regard Data Access as part of User Interaction.   This is why


the
  

OpenLink Data Explorer extension (for instance) simply adds the


ability
  

to View | Linked Data Sources to its host browser.  All it is really
doing is providing the user with a route to a beachead from which a
query could be beamed (i.e. SPARQL + Full Text under the covers across
the data sources in it's history).

Give users the  option to use their own cognitive skills to interact
with the data (e.g. pivot, project, and traverse based on their own
specific needs).

We are repeating a long history of not understanding intersection of
User Interaction and User Cognitive Skills if we don't provide
routes to the data behind subject projections.  This matter has been
the
same since advent of the IT era:

1. Executive Information Systems (what used to be called EIS)
2. Report Writing Systems (Cognos, Crystal Repors, Business Objects

One Last Thing: freebase parallax: user interface for browsing graphs of data

2008-08-17 Thread Kingsley Idehen


David Huynh wrote:


Kingsley Idehen wrote:
Question: Is there any fundamental reason why you cannot expose URIs 
or URLs where you have javascript:{}? Would this break your work in 
anyway?  
Just for you, Kingsley, I have fixed it :-) You might need to 
shift-reload to get the latest code.


I wasn't actually expecting such an intense reaction to just 
javascript:{}. I wonder if that might put off newcomers, who believe 
that the slightest profanity against The URIs on this mailing list 
will always trigger such adverse reactions.


If you expose the URIs, I would then be able to demonstrate what I 
mean using your nice UI, how about that?
Now that that's all behind us, I'm looking forward to see what you 
mean. Please, show us what you've got!


Best,

David




David,

In the nice to have bucket, is is possible for you to use link 
rel=dc:source title=Data Sources type=application/ [atom+xml] | 
[rss+xml] | [rdf+xml]  href=feed-information-resource-url / to 
expose the list of Freebase URLs in you pages. Thus, a list of all URLs 
(once all javascript:{} have been replaced).


Example:
http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president 



Has URLs (instead of: javascript:{}) for each page about a President, 
but this isn't the case re. Filter Results DIV.  Ideally, all the 
Freebase URLs could be exposed via a Feed which lists all the URLs (as 
per the link /  suggestion above, with dc:source as the DC term for 
data sources).


So one last thing, which cannot be expensive to implement since you are 
simply substituting a few more javascript:{} for URLs and the listing 
them is a data sources collection resource (using Atom, RSS 2.0, or RSS 
1.0 or even OPML).


When all of this is in place, I will then have multiple points from 
which to launch a SPARQL Query without the user writing a line of SPARQL 
or seeing anyting to do with RDF. They will simply feel the FORCE of the 
Linked Data :-)



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: freebase parallax: user interface for browsing graphs of data

2008-08-18 Thread Kingsley Idehen


Dan Brickley wrote:

David Huynh wrote:


Kingsley Idehen wrote:
Question: Is there any fundamental reason why you cannot expose URIs 
or URLs where you have javascript:{}? Would this break your work 
in anyway?  
Just for you, Kingsley, I have fixed it :-) You might need to 
shift-reload to get the latest code.


I wasn't actually expecting such an intense reaction to just 
javascript:{}. I wonder if that might put off newcomers, who 
believe that the slightest profanity against The URIs on this mailing 
list will always trigger such adverse reactions.


Ah, we're an excitable bunch around here ;)

Regardless of whether these data URLs are available, the UI work is 
exciting and I'm glad you shared it with the W3C lists. If others have 
smart ideas for visualising and navigating RDFesque datasets (whether 
closed or open by various definitions), let me be clear: they're very 
welcome to post them to [EMAIL PROTECTED] I hope our enthusiasm 
hereabouts for open data doesn't discourage people for sharing 
innovative UI ideas. We need all the help we can get! :)



Dan,

I hope you understand my response to David was very much in the vein of 
two-for-one by taking what was a private discussion (between David and 
I) public for broader knowledge exchange and general discourse purposes.


In similar vein, it's well worth SW people poking around 
http://code.google.com/apis/visualization/
http://processing.org/ and similar, ... the fact that these don't 
(yet) natively support RDF is no reason not to learn from them...


On my part I see User Interaction as the prime issue, with Data 
Access being part of  User Interaction.


Once David complete the tweaks I've requested, I believe everything will 
come together in pretty obvious way. :-)


Kingsley



cheers,

Dan

--
http://danbri.org/





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: One Last Thing: freebase parallax: user interface for browsing graphs of data

2008-08-19 Thread Kingsley Idehen


David Huynh wrote:

Kingsley Idehen wrote:

David,

In the nice to have bucket, is is possible for you to use link 
rel=dc:source title=Data Sources type=application/ [atom+xml] | 
[rss+xml] | [rdf+xml]  href=feed-information-resource-url / to 
expose the list of Freebase URLs in you pages. Thus, a list of all 
URLs (once all javascript:{} have been replaced).


Example:
http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president 



Has URLs (instead of: javascript:{}) for each page about a 
President, but this isn't the case re. Filter Results DIV.  
Ideally, all the Freebase URLs could be exposed via a Feed which 
lists all the URLs (as per the link /  suggestion above, with 
dc:source as the DC term for data sources).


So one last thing, which cannot be expensive to implement since you 
are simply substituting a few more javascript:{} for URLs and the 
listing them is a data sources collection resource (using Atom, RSS 
2.0, or RSS 1.0 or even OPML).


When all of this is in place, I will then have multiple points from 
which to launch a SPARQL Query without the user writing a line of 
SPARQL or seeing anyting to do with RDF. They will simply feel the 
FORCE of the Linked Data :-)
It's actually not trivial to add that link rel=dc:source / as 
Parallax is a dynamic mostly client-side application that has only one 
URL, unless you explicitly ask for permanent links. It's like Google 
Maps.


So, ball is in your court again. :-) There are plenty of proper web 
links now to work with. Let's see some force :-) It'd be nice to have 
a screencast of that, too, so it's easy to follow and forward.


David


David,

Okay, will meet you half way for sure.

Ball back in my court, and screencast already assumed :-)



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: freebase parallax: user interface for browsing graphs of data

2008-08-20 Thread Kingsley Idehen


Tim Berners-Lee wrote:



On 2008-08 -19, at 21:27, David Huynh wrote:



The trouble is of course when the whole web is the database, it's 
hard to suggest those relationships (connections) for a set of entities.
How might one solve that problem? I suppose something like Swoogle 
can help. Is that what Tabulator uses to know what data is on the SW?


Swoogle? Centralized index? No, not at all. The Tabulator is a browser 
for linked data.
The convention for linked data is that if I do a GET on the URI 
identifying something, the returned information will include the 
incoming and outgoing links that a reasonable person might be 
interested in following.  Linked data clients dereference any URIs 
they have not come across before, and pick up either data directly or 
clues as to where to find what more data.  So at each point, you know 
what the options are leading on.


When you pick up a set of related things, of course, there will be 
some presidents who have children and some who don't.  And there will 
be some presidents which a have bunch of obscure properties. 
Especially once you have overlaid data from many sources.  So then 
there may be a case for having lenses linked from ontologies to allow 
one for example to focus on geneology or political career.


It gets more complicated when you try to automatically build an 
interface for allowing people to input data about a president, chose 
which fields to offer.


I'd like to see the Freebase data as linked data on the web ... then 
we could try all our other UIs on the same data!  What would it take 
to put in some shim interface to Freebase?


Tim

PS:  There are places where a centralized index will help though, like 
finding people's FOAF page from their email address.









David,

Which brings us back to the initial point: expose the URIs (you've done 
this partially) to Linked Data aware agents :-)  This, as I've already 
indicated, comes down to:


1. Replace all javascript:{} with their actual Freebase URIs (which 
already exist)
2. Use an information resource to collate all the URIs that are in scope 
at a given point in time in you UI, and then expose the resource URI via 
link rel=dc:source ../ (you can use RSS, Atom, OPML for this, the 
key thing is to list the URIs)


By doing the above, your Visualization oriented information resource and 
the data sources behind remain Linked Data Web accessible.


Anyway, you've already met me half way (re. yesterday's comments), so 
I'll crack on with my bit: showing how your interface can provide a 
beachead for launching a SPARQL Full Text query (without the user 
writing any SPARQL or seeing any RDF in plain sight).



Tim et al: 
There is a first cut of the last but one Freebase dump in Linked Data 
form (to the degree attainable) at:
http://linkeddata.openlinksw.com:8891/sparql (the graph URI for SPARQL 
FROM is : http://linkeddata.openlinksw.com/freebase# )


We are working with Francois Belleau (of http://bio2rdf.org) to produce 
an updated release, but even that release is going to require some work 
in order for the graph to be totally coherent (*a long story*).


Also note, we are RDFizing Freebase on the fly and producing slightly 
better Linked Data, so if I go to 
http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president 
and simply do a View | Linked Data Sources via the ODE XPI's context 
menu against a President's Freebase URI I get: http://tinyurl.com/6pxetc .



To conclude, we have Linked Data in two forms for the Freebase Data Space:

1. Their Quad Dumps transformed into RDF and then loaded into Virtuoso 
resulting in a SPARQL endpoint and a Linked Data Space
2. Linked Data generated on the Fly via our RDFization cartridges with 
proxy URIs used as mechanism for endowing the resulting Linked Data 
Space with dereferencable URIs


Tim: Tabulator should work with either.

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)

2008-08-20 Thread Kingsley Idehen


Aldo Bucchi wrote:

HI,

Scanning the thread on Parallax I see some terms reoccurring:

Outgoing Connections.
Lenses.
Lists.
Facets.
Free text search.
IFP to IRI resolution.
Find documents that contain IRIs
etc...

They are all implemented in different ways but tend to share semantics
across different browsers and services.
How far are we from defining a modular framework so we can mix and
math these as atomic interaction pieces?

Both services and probably UI parts.

Of course, RDF and HTTP can be used to describe them and deliver the
descriptions and, in the case of widgets, some OOTB implementations.
XForms, Dojo widgets, SWFs?

I have done something similar but much simpler in a Flex platform ( I
serve Flex modules, described in RDF and referenced by Fresnel vocabs,
but only for presentation ).
And then on a functional side I have several services that do
different things, and I can hot swap them.
For example, the free text search service is a (S)WS.
Faceter service idem.

I guess we still need to see some more diversity to derive a taxonomy
and start working on the framework.
But it is nice to keep this in sight.

The recurring topics.

Best,
A


  

Aldo,

Really nice to see you are looking at things holistically.

As you can see, we are veering gradually towards recognizing that the 
Web, courtesy of HTTP,  gives us a really interesting infrasructure 
for the time-tested MVC pattern (I've been trying to bring attention to 
this aspect of the Web for a while now).


If you look at ODE (closely) you will notice it's an MVC vessel. We have 
components for Data Access (RDFiztion Cartridges), components for UI 
(xslt+css templates and fresnel+xslt+css templates), and components for 
actions (*Cartridges not released yet*).


We've tried to focus on the foundation infrastructure that uses HTTP for 
the messaging across M-V-C so that you get:

M--http-V---http---C

Unfortunately, our focus on the MC doesn't permeate. Instead, we find 
all focus coming at us from the V part where we've released minimal 
templates with hope that 3rd parties will eventually focus on Display 
Cartridges (via Fresnel, XSLT+SPARQL, xml+xslt+css, etc..).


btw - David Schwabe [1] also alluded to the architectural modularity 
that I am fundamentally trying to bring to broader attention in this 
evolving  conversation re.  Linked oriented Web applications.


The ultimate goal is to deliver a set of options that enable Web Users 
to Explore the Web coherently and productively (imho).


Humans can only do so much, and likewise Machines, put both together and 
we fashion a recipe for real collective intelligence (beyond  the 
buzzword).  We desperately need to tap into collective intelligenceen 
route to solving many of the real problems facing the world today.


The Web should make seeing and connecting the dots easier, but this is 
down to the MVC combo as opposed to any single component of the pattern  :-)



Links:

1. http://lists.w3.org/Archives/Public/public-lod/2008Aug/att-0106/00-part

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Little Correction: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)

2008-08-20 Thread Kingsley Idehen


[snip]



btw - David Schwabe [1] also alluded to the architectural modularity 
that I am fundamentally trying to bring to broader attention in this 
evolving  conversation re.  Linked oriented Web applications.

Aldo,

In my earlier post, excerpted above, I incorrectly referred to Daniel 
Schwabe as David Schwabe.



This post  is my s/David/Daniel style correction, for the record, on our 
sticky Web :-)


The ultimate goal is to deliver a set of options that enable Web Users 
to Explore the Web coherently and productively (imho).


Humans can only do so much, and likewise Machines, put both together 
and we fashion a recipe for real collective intelligence (beyond  
the buzzword).  We desperately need to tap into collective 
intelligenceen route to solving many of the real problems facing the 
world today.


The Web should make seeing and connecting the dots easier, but this is 
down to the MVC combo as opposed to any single component of the 
pattern  :-)



Links:

1. 
http://lists.w3.org/Archives/Public/public-lod/2008Aug/att-0106/00-part





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: soliciting your favorite (public) SPARQL queries!

2008-08-21 Thread Kingsley Idehen


Lee Feigenbaum wrote:


Hi everyone,

I'm putting together a SPARQL by Example tutorial, which is, as the 
name suggests, a step-by-step introduction to SPARQL taught almost 
entirely through complete, runnable SPARQL queries.


So far, I've gathered a great deal of example queries myself, but I 
know that many subscribers to these lists probably have favorite 
queries of their own that you might be willing to share with me.


I'm looking for:

1) SPARQL queries
2) ...that can be run by anyone (no private data sets)
3a)...either by running the query against a public SPARQL endpoint
3b)...or by using a public SPARQL endpoint that will fetch 
HTTP-accessible RDF data (e.g. sparql.org or demo.openlinksw.com)

4) ...that answers a real* question
5) ...and that is fun!**

* real is in the eye of the beholder, I imagine, but I'm not looking 
for  finds the predicates that relate ex:s and ex:o in this sample 
RDF graph


** fun is also in the eye of the beholder. fun can be a query on fun 
data; a clever query that may illustrate a particular SPARQL construct 
(trick); a query that integrates interesting information; a query 
with surprising results; etc.


thanks to anyone who is able to contribute!
Lee

PS I plan to make the tutorial slides available online under an 
appropriate CC license once they are completed.




Lee,

Here are our contributions (endpoint: http://demo.openlinksw.com/sparql):

1. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/NorthWindREF  
- against live RDF Views of Northwind SQL Database (Products, Customers, 
Orders etc.)


2. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSSIOCRef - 
collection of SPARQL queries against SIOC based exploration of OpenLink 
Data Space s blogs, wikis, bookmarks, calendars, feed subscriptions etc..)


3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSFOAFRef - 
collection of SPARQL queries for FOAF navigation of OpenLink Data Spaces


4. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSSKOSRef - 
collection of SPARQL queries for SKOS based exploration of Tag Spaces 
associated with OpenLink Data Spaces


5. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/SIOCRefDocs- 
collection of SPARQL queries using SIOC to explore Virtuoso's Online 
Docs Data Space


6. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/WordPressSIOCRef 
- SPARQL queries for exploring Wordpress Data when Deployed via 
Virtuoso's PHP runtime hosting


7.  
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/MediaWikiSIOCRef 
- ditto re. MediaWiki


8. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/PHPBB3SIOCRef 
- ditto phpBB


9. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/DrupalSIOCRef 
- ditto Drupal


10. http://www4.wiwiss.fu-berlin.de/benchmarks-200801/#queries - DBpedia 
Benchmark Queries (* run against the DBpedia SPARQL endpoint at: 
http://dbpedia.org/sparql *)


11. http://tinyurl.com/5fjp5h - Collection of Inference based Queries 
against DBpedia that leverage the Yago Class Hierarchy


Note: re. the ODS examples (items 2 - 4 ), if you change the SPARQL 
endpoint to: http://community.linkeddata.org/sparql , and Graph IRI in 
the SPARQL FROM clause to: http://community.linkeddata.org/sparql, you 
can then run the SIOC, FOAF, SKOS examples against the live LOD Data 
Space with more interesting results.



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)

2008-08-21 Thread Kingsley Idehen


Aldo Bucchi wrote:

Hello,

( replies inlined )

On Wed, Aug 20, 2008 at 9:19 PM, Kingsley Idehen [EMAIL PROTECTED] wrote:
  

Aldo Bucchi wrote:


HI,

Scanning the thread on Parallax I see some terms reoccurring:

Outgoing Connections.
Lenses.
Lists.
Facets.
Free text search.
IFP to IRI resolution.
Find documents that contain IRIs
etc...

They are all implemented in different ways but tend to share semantics
across different browsers and services.
How far are we from defining a modular framework so we can mix and
math these as atomic interaction pieces?

Both services and probably UI parts.

Of course, RDF and HTTP can be used to describe them and deliver the
descriptions and, in the case of widgets, some OOTB implementations.
XForms, Dojo widgets, SWFs?

I have done something similar but much simpler in a Flex platform ( I
serve Flex modules, described in RDF and referenced by Fresnel vocabs,
but only for presentation ).
And then on a functional side I have several services that do
different things, and I can hot swap them.
For example, the free text search service is a (S)WS.
Faceter service idem.

I guess we still need to see some more diversity to derive a taxonomy
and start working on the framework.
But it is nice to keep this in sight.

The recurring topics.

Best,
A



  

Aldo,

Really nice to see you are looking at things holistically.



I showed up with a narrow interface, I know ;)

  

As you can see, we are veering gradually towards recognizing that the Web,
courtesy of HTTP,  gives us a really interesting infrasructure for the
time-tested MVC pattern (I've been trying to bring attention to this aspect
of the Web for a while now).

If you look at ODE (closely) you will notice it's an MVC vessel. We have
components for Data Access (RDFiztion Cartridges), components for UI
(xslt+css templates and fresnel+xslt+css templates), and components for
actions (*Cartridges not released yet*).



Ah... I remember telling Daniel Lewis something was missing from his
UPnP diagram: a way to modify the Model.
aka: a Controller / Actions.

You are right, technically an agent like ODE ( assuming you can hook
in actions ) is all that you need to allow users to interact with
linked data.

Let's say that this sort of solution can cover 80% of user interaction
cases ( launching simple actions and direct manipulation of resources
), and operates on top of 80% of data ( anything that can be published
as linked data/SPARQL and fits within the expressiveness of RDF's
abstract model ).
Not a bad MVC structure at all!

So, how do you plan on hooking up the actions to the shell, is this
in the cartridges?
How will they surface. Context menu?
  
Everything lives in a REST or SOAP accessible Cartridge. ODE just talks 
REST or SOAP .


For instance, ODE uses REST calls to the Sponger Service when RDFizing, 
but it can just as well use Triplr. We've just put out a new ODE release 
with an improved Preferences dialog that makes the mixing and matching 
of Renderers and RDFizers clearer.


Re. Display Cartridges (Fresnel Templates) the same would apply, but in 
this case we just deal with the URIs of the templates. Ditto in the case 
of Actions.


URIs and REST are all that we need fundamentally, which is just about 
leveraging what the Web has always offered.
  

We've tried to focus on the foundation infrastructure that uses HTTP for the
messaging across M-V-C so that you get:
M--http-V---http---C

Unfortunately, our focus on the MC doesn't permeate. Instead, we find all
focus coming at us from the V part where we've released minimal templates
with hope that 3rd parties will eventually focus on Display Cartridges (via
Fresnel, XSLT+SPARQL, xml+xslt+css, etc..).



Well. The M part is the data, isn't it? ( so it is permeating, people
are publishing data ).
  


Yes.

Unless you mean building some higher functionality services ( on top
of SPARQL and RDF ) such as faceters, free text search, IFP
resolution, etc. But in that case it is also moving forward, although
not with a standardized interface.
This could be thought of as higher level Data Access components.

  
Correct, with Linked Data at the base. This is why I always refere to 
Linked Data as the foundation layer of the Semantic Web innovation 
continuum.



The C part... that's another story.
As I pointed out before, you need to define the way and an environment
to hook in the actions. What is the shell?
  
ODE is an inner shell / enclave  example, but in reality the Web is the 
outer shell (as long as you honor links and link dereferencing). What is 
missing is a vocabulary for actions.


Example: if you combined GoodRelations Ontology and Web Services part of 
SIOC (the extended modules part)  you get a Linked Data Space that not 
only describes a service vendor, but also one that formally describes 
how to consummate a transaction with said vendor, with granularity at 
the service level (so two services with different signatures from

Re: soliciting your favorite (public) SPARQL queries!

2008-08-21 Thread Kingsley Idehen
 and ex:o
in this sample RDF graph

** fun is also in the eye of the beholder. fun can be a query
on fun data; a clever query that may illustrate a particular
SPARQL construct (trick); a query that integrates
interesting information; a query with surprising results; etc.

thanks to anyone who is able to contribute!
Lee

PS I plan to make the tutorial slides available online under
an appropriate CC license once they are completed.



-- 
Dr. Axel Polleres, Digital Enterprise Research Institute (DERI)

email: [EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
 url: http://www.polleres.net/

Everything is possible:
rdfs:subClassOf rdfs:subPropertyOf rdfs:Resource.
rdfs:subClassOf rdfs:subPropertyOf rdfs:subPropertyOf.
rdf:type rdfs:subPropertyOf rdfs:subClassOf.
rdfs:subClassOf rdf:type owl:SymmetricProperty.



Adrian,

Re. SPARQL  Aggregates, see: 
http://esw.w3.org/topic/SPARQL/Extensions/Aggregates



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Mozilla Labs Ubiquity and Linked Data

2008-09-04 Thread Kingsley Idehen


Hausenblas, Michael wrote:

UPDATE: After a rather short night (I tell you, toying around with
ubiquity is *really* addictive) there are now two more commands
available: one for querying DBpedia (simple skos:subject on categories)
and a curl HEAD command (should be self-explanatory).

If you develop (Web of Data) commands as well, please consider
announcing them on the Mozilla Wiki [1].

Cheers,
Michael

[1]
https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data

--
 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
  
 http://www.joanneum.at/iis/

--
 
  


Here is our initial Ubiquity command for Linked Data which has the 
command pattern: describe-page URL:

http://demo.openlinksw.com/ubiq/

Of course there is more to come re. a generic SPARQL command etc..


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Correction: Mozilla Labs Ubiquity and Linked Data

2008-09-04 Thread Kingsley Idehen


Kingsley Idehen wrote:

Hausenblas, Michael wrote:

UPDATE: After a rather short night (I tell you, toying around with
ubiquity is *really* addictive) there are now two more commands
available: one for querying DBpedia (simple skos:subject on categories)
and a curl HEAD command (should be self-explanatory).

If you develop (Web of Data) commands as well, please consider
announcing them on the Mozilla Wiki [1].

Cheers,
Michael

[1]
https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data

--
 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
   http://www.joanneum.at/iis/
--
 
  


Here is our initial Ubiquity command for Linked Data which has the 
command pattern: describe-page URL:

http://demo.openlinksw.com/ubiq/


Command is currently: page-describe URL (as per index.html), but it 
will ultimately be: describe-page URL :-)


Kingsley


Of course there is more to come re. a generic SPARQL command etc..





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Mozilla Labs Ubiquity and Linked Data

2008-09-04 Thread Kingsley Idehen


Hausenblas, Michael wrote:

Kingsley,
 
Great work! Again, OpenLink shows that it is not only a good but also a very quick player in the linked data field. If I only had the time now (/me currently at TRIPLE-I conference and busy applauding Tom ;) I'd have done it myself, sorry. Anyways, good to see the uptake. 
 
Two minor comments: (i) the command, I guess actually is 'page-describe' and (ii) we might wanna make it configurable to play with other SW browsers such as Tabulator (as you proposed, IIRC). 
  

Michael,

I think it does play well with the other Linked Data clients (see: page 
footer) :-)



Kingsley
 
Cheers,

Michael
 
--

 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
 Steyrergasse 17, A-8010 Graz, AUSTRIA

 office
phone: +43-316-876-1193 (fax:-1191)  
   e-mail: [EMAIL PROTECTED]
  web: http://www.joanneum.at/iis/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.joanneum.at/iis/ 


 private
   mobile: +43-660-7621761
  web: http://www.sw-app.org/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.sw-app.org/ 
--





From: [EMAIL PROTECTED] on behalf of Kingsley Idehen
Sent: Thu 2008-09-04 17:20
To: public-lod@w3.org
Cc: [EMAIL PROTECTED]; public-lod@w3.org
Subject: Re: Mozilla Labs Ubiquity and Linked Data




Hausenblas, Michael wrote:
  

UPDATE: After a rather short night (I tell you, toying around with
ubiquity is *really* addictive) there are now two more commands
available: one for querying DBpedia (simple skos:subject on categories)
and a curl HEAD command (should be self-explanatory).

If you develop (Web of Data) commands as well, please consider
announcing them on the Mozilla Wiki [1].

Cheers,
  Michael

[1]
https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data

--
 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
 
 http://www.joanneum.at/iis/

--

 



Here is our initial Ubiquity command for Linked Data which has the
command pattern: describe-page URL:
http://demo.openlinksw.com/ubiq/

Of course there is more to come re. a generic SPARQL command etc..


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com http://www.openlinksw.com/ 









  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Mozilla Labs Ubiquity and Linked Data Other Linked Data Clients

2008-09-04 Thread Kingsley Idehen


Hausenblas, Michael wrote:

Kingsley,
 
Great work! Again, OpenLink shows that it is not only a good but also a very quick player in the linked data field. If I only had the time now (/me currently at TRIPLE-I conference and busy applauding Tom ;) I'd have done it myself, sorry. Anyways, good to see the uptake. 
 
Two minor comments: (i) the command, I guess actually is 'page-describe' and (ii) we might wanna make it configurable to play with other SW browsers such as Tabulator (as you proposed, IIRC). 
 
Cheers,

Michael
 
--

 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
 Steyrergasse 17, A-8010 Graz, AUSTRIA

 office
phone: +43-316-876-1193 (fax:-1191)  
   e-mail: [EMAIL PROTECTED]
  web: http://www.joanneum.at/iis/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.joanneum.at/iis/ 


 private
   mobile: +43-660-7621761
  web: http://www.sw-app.org/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.sw-app.org/ 
--





From: [EMAIL PROTECTED] on behalf of Kingsley Idehen
Sent: Thu 2008-09-04 17:20
To: public-lod@w3.org
Cc: [EMAIL PROTECTED]; public-lod@w3.org
Subject: Re: Mozilla Labs Ubiquity and Linked Data




Hausenblas, Michael wrote:
  

UPDATE: After a rather short night (I tell you, toying around with
ubiquity is *really* addictive) there are now two more commands
available: one for querying DBpedia (simple skos:subject on categories)
and a curl HEAD command (should be self-explanatory).

If you develop (Web of Data) commands as well, please consider
announcing them on the Mozilla Wiki [1].

Cheers,
  Michael

[1]
https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data

--
 Michael Hausenblas, MSc.
 Institute of Information Systems  Information Management
 JOANNEUM RESEARCH Forschungsgesellschaft mbH
 
 http://www.joanneum.at/iis/

--

 



Here is our initial Ubiquity command for Linked Data which has the
command pattern: describe-page URL:
http://demo.openlinksw.com/ubiq/

Of course there is more to come re. a generic SPARQL command etc..


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com http://www.openlinksw.com/ 









  

Michael,

I just realized that you may have been requesting that we use our Proxy 
URIs in the calls we make to the other Linked Data Clients (in the page 
footer), which is how we feed them units of RDFization via our 
Proxy/Wrapper URIs. That's certainly worth doing.


There will be an update with this feature. 



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Announce: OpenLink Data Explorer Extension Update

2008-09-06 Thread Kingsley Idehen


All,

We are please to announce the availability of an update release of the 
OpenLink Data Explorer (ODE) extension for Firefox [1][2].


Features and enhancements include:

1. New Page Description main and context menu item that provides a an 
(X)HTML representation of a resource description
2. First look at new Find feature which replaces Search and seeks to 
demonstrate that Linked Data helps us Find rather than Search for 
stuff (more to come re. this powerful feature in follow-on releases)

3. A number of aesthetic fixes

We've also updated the ubiquity commands [3] that provide commandline 
access points to the same services that drive ODE.


Links:

1. https://addons.mozilla.org/en-US/firefox/addon/8062
2. http://ode.openlinksw.com/#Download
3. http://demo.openlinksw.com/ubiq (*read my most recent blog post re. 
subscribe and unsubscribe howto*)


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








World Bank API

2008-09-06 Thread Kingsley Idehen


All,

For those unaware of this development, the World Bank has now released 
APIs for their data space on the Web [1].



Links:

1. http://developer.worldbank.org/

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: New Calais proxy could grow Linked Data Cloud

2008-09-23 Thread Kingsley Idehen


Paul Miller wrote:
Members of this list might be interested in my write-up of 
ThomsonReuters' latest beta service... which I think will prove pretty 
useful in growing the Linked Data cloud... especially for news content 
from the BBC et al...


http://blogs.zdnet.com/semantic-web/?p=194

Paul

--
Paul Miller
Technology Evangelist, Talis
w: www.talis.com/ http://www.talis.com/  skype: napm1971
mobile/cell: +44 7769 740083

http://blogs.zdnet.com/semantic-web/

_www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er_





Paul,

How does this actually benefit or contribute to the Linked Data Cloud? I 
ask specifically because URIs (of the dereferencable variety) are 
missing in action. Hopefully, I am completely overlooking something here :-)


We tend to use the term Proxy or Wrapper to describe solutions in 
the Linked Data realm that generate dereferencable URIs based RDF 
graphs  (aka. Linked Data Spaces) on the fly via RDF-ization middleware.


If possible, please encourage the OpenCalais folks (Tom et al.) to 
respond to my comments above via a response to this post.


Example Proxy / Wrapper URIs in the wild:

1. 
http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/abraham_lincoln 
- Document about Abraham Lincoln
2. 
http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.freebase.com/view/en/abraham_lincoln%23this  
- Abraham Lincoln the Entity of type foaf:Person that is also a sioc:Item
3. 
http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.crunchbase.com/company/thomson-reuters%23this 
- Thompson Reuters the Entity of type foaf:Organization that is also a 
sioc:Item
4. http://www4.wiwiss.fu-berlin.de/flickrwrappr/photos/Thomson_Reuters - 
Thompson Reuters photos from Flickr
5. 
http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fwww4.wiwiss.fu-berlin.de%2Fflickrwrappr%2Fphotos%2FThomson_Reuters 
- Browser view of the data space exposed by the Flickr wrapper URI




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Freebase wrapper

2008-09-23 Thread Kingsley Idehen


Georgi Kobilarov wrote:

Hi Kingsley,

  

Example Proxy / Wrapper URIs in the wild:

1.



http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/a
  

braham_lincoln
- Document about Abraham Lincoln
2.



http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/
  

rdf/http://www.freebase.com/view/en/abraham_lincoln%23this
- Abraham Lincoln the Entity of type foaf:Person that is also a
sioc:Item



these URIs unfortunately do not show any data.
Would be great to see your freebase wrapper in action...

Best,
Georgi
  

Georgi,

As per usual, I announce a demo at a time when maintenance occurs on the 
demo server :-(


See: 
http://demo.openlinksw.com/about/html/http://www.freebase.com/view/en/abraham_lincoln


Kingsley

--
Georgi Kobilarov
Freie Universität Berlin
www.georgikobilarov.com


  

-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On
Behalf Of Kingsley Idehen
Sent: Tuesday, September 23, 2008 2:02 PM
To: Paul Miller
Cc: public-lod@w3.org
Subject: Re: New Calais proxy could grow Linked Data Cloud


Paul Miller wrote:


Members of this list might be interested in my write-up of
ThomsonReuters' latest beta service... which I think will prove
  

pretty


useful in growing the Linked Data cloud... especially for news
  

content


from the BBC et al...

http://blogs.zdnet.com/semantic-web/?p=194

Paul

--
Paul Miller
Technology Evangelist, Talis
w: www.talis.com/ http://www.talis.com/  skype: napm1971
mobile/cell: +44 7769 740083

http://blogs.zdnet.com/semantic-web/

_www.linkedin.com/in/pau1mi11er
  

http://www.linkedin.com/in/pau1mi11er_




  

Paul,

How does this actually benefit or contribute to the Linked Data Cloud?
I
ask specifically because URIs (of the dereferencable variety) are
missing in action. Hopefully, I am completely overlooking something
here :-)

We tend to use the term Proxy or Wrapper to describe solutions in
the Linked Data realm that generate dereferencable URIs based RDF
graphs  (aka. Linked Data Spaces) on the fly via RDF-ization
middleware.

If possible, please encourage the OpenCalais folks (Tom et al.) to
respond to my comments above via a response to this post.

Example Proxy / Wrapper URIs in the wild:

1.



http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/a
  

braham_lincoln
- Document about Abraham Lincoln
2.



http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/
  

rdf/http://www.freebase.com/view/en/abraham_lincoln%23this
- Abraham Lincoln the Entity of type foaf:Person that is also a
sioc:Item
3.



http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/
  

rdf/http://www.crunchbase.com/company/thomson-reuters%23this
- Thompson Reuters the Entity of type foaf:Organization that is also a
sioc:Item
4. http://www4.wiwiss.fu-berlin.de/flickrwrappr/photos/Thomson_Reuters
-
Thompson Reuters photos from Flickr
5.



http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fwww4.wiwiss.fu
  

-berlin.de%2Fflickrwrappr%2Fphotos%2FThomson_Reuters
- Browser view of the data space exposed by the Flickr wrapper URI



--


Regards,

Kingsley Idehen   Weblog:


http://www.openlinksw.com/blog/~kidehen
  

President  CEO
OpenLink Software Web: http://www.openlinksw.com







  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: New Calais proxy could grow Linked Data Cloud

2008-09-23 Thread Kingsley Idehen


Thomas Tague wrote:

LOD Group:

Tom,

First off, welcome!

Really happy to see you've taken the time engage the LOD community re. 
your team's efforts and long term objectives.


Questions follow inline below.


First, a philosophical point and then a few facts. 

When your child first learns to read you don't discard that because 
they haven't yet graduated from college. You know college is coming, 
you're already thinking about college, you may actually be actively 
working on college - but the first words are still important.


Calais is learning to read. We firmly believe in releasing building 
blocks when they become available rather than waiting (and waiting and 
waiting) for the entire solution to be ready.


A few specific facts to make it clearer where SemanticProxy fits in:

1) We will have de-referenceable URIs for every entity extracted by 
Calais by the end of this year. The engineering is done and we're in 
active design and build mode. We haven't finished the analysis yet - 
but this will be millions of endpoints on the day we go live.
Please clarify what you mean by endpoints.  Over here it might refer 
to SPARQL endpoints or derferencable URIs. I suspect you mean URIs, but 
clarification from you will aid others.


2) A *subset* of those entity types will absolutely have links to 
other linked data sources when we go live. Right now we know there 
will be substantive links for companies, geographies and a few of the 
easy ones like music, books, etc. We'll expand on that set over time 
and have a goal of setting up a community-based mechanism for 
enhancing the links over time.


Will linkage apply to instance data and associated definitions data 
(ontology / schema / data dictionary) for the Thompson Reuters linked 
data spaces?
Will you be using shared ontologies where such exist, or at the very 
least put out your ontology in RDFS or OWL?  Even doing this  open up 
the doors for community participation in the data definitions linkage 
effort (e.g. what's happened re. UMBEL, OpenCyc, Yago, and Wordnet).


3) At the end of this month (September) as part of Release 3.1 we'll 
be releasing company and geography disambiguation as a component of 
the metadata generation process. The company disambiguation is based 
on a lexicon of over 16M company aliases + additional hinting and we 
have a similar approach with geography. 

Great news, but the real utility of such work will always be easier to 
imbibe, by this community in particular,  if the resulting output is an 
RDF Linked Data Space rather than an RDF Data Island :-)


Again, great to have you outline your development road-map here, I 
certainly believe this will ultimately be a great contribution to the 
burgeoning Linked Data Web.



Kingsley


Question? Ideas? Fire away.

Tom


On Tue, Sep 23, 2008 at 8:21 AM, Paul Miller [EMAIL PROTECTED] 
mailto:[EMAIL PROTECTED] wrote:


From the post...

SemanticProxy will return dereferenceable Linked Data URIs by the
end of this quarter.

Paul

--
Paul Miller
Technology Evangelist, Talis
w: www.talis.com/ http://www.talis.com/  skype: napm1971
mobile/cell: +44 7769 740083

http://blogs.zdnet.com/semantic-web/

_www.linkedin.com/in/pau1mi11er
http://www.linkedin.com/in/pau1mi11er_




On 23 Sep 2008, at 13:02, Kingsley Idehen wrote:


Paul Miller wrote:

Members of this list might be interested in my write-up of
ThomsonReuters' latest beta service... which I think will prove
pretty useful in growing the Linked Data cloud... especially for
news content from the BBC et al...

http://blogs.zdnet.com/semantic-web/?p=194

Paul

--
Paul Miller
Technology Evangelist, Talis
w: www.talis.com/ http://www.talis.com/
http://www.talis.com/  skype: napm1971
mobile/cell: +44 7769 740083

http://blogs.zdnet.com/semantic-web/

_www.linkedin.com/in/pau1mi11er
http://www.linkedin.com/in/pau1mi11er
http://www.linkedin.com/in/pau1mi11er_





Paul,

How does this actually benefit or contribute to the Linked Data
Cloud? I ask specifically because URIs (of the dereferencable
variety) are missing in action. Hopefully, I am completely
overlooking something here :-)

We tend to use the term Proxy or Wrapper to describe
solutions in the Linked Data realm that generate dereferencable
URIs based RDF graphs  (aka. Linked Data Spaces) on the fly via
RDF-ization middleware.

If possible, please encourage the OpenCalais folks (Tom et al.)
to respond to my comments above via a response to this post.

Example Proxy / Wrapper URIs in the wild:

1.

http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/abraham_lincoln
- Document about Abraham Lincoln
2.

http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.freebase.com/view/en/abraham_lincoln%23this
 - Abraham Lincoln the Entity of type

Re: New Calais proxy could grow Linked Data Cloud

2008-09-23 Thread Kingsley Idehen


Tom Heath wrote:

Hi Chris,

2008/9/23 Chris Sizemore [EMAIL PROTECTED]:
  

does anyone else expect/hope that the Linked Data part of dereferenceable
Linked Data URIs  implies that these URIs will be linked up to, for
instance, dbPedia, umbel, musicbrainz, etc etc? otherwise it wouldn't really
be linked data per se, no?



I hope that that's the case, although I don't know if it will turn out
that way. It may be a case of Calais confusing Linked Data with
linkable data. It is certainly my expectation (and that of the entire
LD community, I would argue) that Linked Data does actually include
those links, in case anyone was in any doubt ;)

Tom.


  

Chris / Tom,

I would decompose Linked Data, in this context, as follows:

1. URIs can be used to de-reference an RDF based resource description
2. Resource description includes de-rferencable URIs across Subject, 
Predicate, and Object (where applicable) in constituent triples


Linkage to other Linked Data spaces  (instances and classes) is a 
desired added bonus, but not the essential definition (imho) :-)


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL

2008-09-24 Thread Kingsley Idehen


Paul Gearon wrote:

On Mon, Sep 22, 2008 at 3:47 AM, Eyal Oren [EMAIL PROTECTED] wrote:
  

On 09/19/08/09/08 23:12 +0200, Orri Erling wrote:


Has has there been any analysis on whether there is a *fundamental*
reason for such performance difference? Or is it simply a question of
maturity; in other words, relational db technology has been around for a
very long time and is very mature, whereas RDF implementations are still
quite recent, so this gap will surely narrow ...?


This is a very complex subject.  I will offer some analysis below, but
this I fear will only raise further questions.  This is not the end of the
road, far from it.
  

As far as I understand, another issue is relevant: this benchmark is
somewhat unfair as the relational stores have one advantage compared to the
native triple stores: the relational data structure is fixed (Products,
Producers, Reviews, etc with given columns), while the triple representation
is generic (arbitrary s,p,o).



This point has an effect on several levels.

For instance, the flexibility afforded by triples means that objects
stored in this structure require processing just to piece it all
together, whereas the RDBMS has already encoded the structure into the
table. Ironically, this is exactly the reason we
(Tucana/Kowari/Mulgara) ended up building an RDF database instead of
building on top of an RDBMS: The flexibility in table structure was
less efficient that a system that just knew it only had to deal with
3 columns. Obviously the shape of the data (among other things)
dictates what it is the better type of storage to use.

A related point is that processing RDF to create an object means you
have to move around a lot in the graph. This could mean a lot of
seeking on disk, while an RDBMS will usually find the entire object in
one place on the disk. And seeks kill performance.

This leads to the operations used to build objects from an RDF store.
A single object often requires the traversal of several statements,
where the object of one statement becomes the subject of the next.
Since the tables are typically represented as
Subject/Predicate/Object, this means that the main table will be
joined against itself. Even RDBMSs are notorious for not doing this
efficiently.

One of the problems with self-joins is that efficient operations like
merge-joins (when they can be identified) will still result in lots of
seeking, since simple iteration on both sides of the join means
seeking around in the same data. Of course, there ARE ways to optimize
some of this, but the various stores are only just starting to get to
these optimizations now.

Relational databases suffer similar problems, but joins are usually
only required for complex structures between different tables, which
can be stored on different spindles. Contrast this to RDF, which needs
to do do many of these joins for all but the simplest of data.

  

One can question whether such flexibility is relevant in practice, and if
so, one may try to extract such structured patterns from data on-the-fly.
Still, it's important to note that we're comparing somewhat different things
here between the relational and the triple representation of the benchmark.



This is why I think it is very important to consider the type of data
being stored before choosing the type of storage to use. For some
applications an RDBMS is going to win hands down every time. For other
applications, an RDF store is definitely the way to go. Understanding
the flexibility and performance constraints of each is important. This
kind of benchmarking helps with that. It also helps identify where RDF
databases need to pick up their act.

Regards,
Paul Gearon


  

Paul,

You make valid points, the problem here is that the benchmark has been 
released without enough clarity about it's prime purpose. To even 
compare RDF Quads Stores with an RDBMS engine when the schema is 
Relational in itself is kinda twisted.


The role of mappers (DR2Q  Virtuoso RDF Views) for instance,  should 
have been made much clearer, maybe in separate results tables. I say 
this because these mappers offer different approaches to projecting 
RDBMS based data in RDF Linked Data form, on the fly, and their purpose 
in this benchmark is all about raw performance and scalability as it 
relates to following RDF Linked Data generation and deployment conditions:


1. Schema is Relational
2. RDF warehouse is impractical

As I am sure you know, we could invert this whole benchmark Open World 
style, and then bring RDBMS engines to their knees by incorporating 
SPARQL query patterns comprised of ?p's and subclasses .


To conclude, the quad store numbers should simply be a conparisons of 
the quad stores themselves, and not the quad stores vs the mappers or 
native SQL. This clarification really needs to make it's way into the 
benchmark narrative.



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO

Re: AW: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL

2008-09-25 Thread Kingsley Idehen


Chris Bizer wrote:

Hi Kingsley and Paul,

Yes, I completely agree with you that different storage solutions fit
different use cases and that one of the main strengths of the RDF data model
is its flexibility and the possibility to mix different schemata.

Nevertheless, it think it is useful to give application developers an
indicator about what performance they can expect when they choose a specific
architecture, which is what the benchmark is trying to do.
  

Chris,

Yes, but the user profile has to be a little clearer. If you separate 
the results in the narrative you achieve the goal. You can use SQL 
numbers as a sort of benchamark if you clearly explain the nature skew 
that SQL enjoys due to the nature of the schema. 

We plan to run the benchmark again in January and it would be great to also
test Tucana/Kowari/Mulgara in this run.

As the performance of RDF stores is constantly improving, let's also hope
that the picture will not look that bad for them anymore then.
  
But at the current time, there is no clear sense of what better means 
:-) What's the goal?


What I fundamentally take from the benchmarks are the following:

1. Native RDF and RDF Views/Mapper scalability is becoming less of an 
issue (of course depending on your choice of product) and we are already 
at the point where this technology can be used for real-world solutions 
that have enterprise level scalability demands and expectations


2. It's impractical to create RDF warehouses from a existing SQL Data 
Sources when you can put RDF Views / Wrappers in front of the SQL Data 
Sources (SQL cost optimization technology has evolved significantly over 
the years across RDBMS engines).



And Yes, I would also like to see Mulgara and others RDF Stores in the 
next round of benchmarks :-)


Kingsley

Cheers,

Chris


-Ursprüngliche Nachricht-
Von: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Im Auftrag
von Kingsley Idehen
Gesendet: Mittwoch, 24. September 2008 20:57
An: Paul Gearon
Cc: [EMAIL PROTECTED]; public-lod@w3.org
Betreff: Re: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena
TDB, D2R Server, and MySQL


Paul Gearon wrote:
  

On Mon, Sep 22, 2008 at 3:47 AM, Eyal Oren [EMAIL PROTECTED] wrote:
  


On 09/19/08/09/08 23:12 +0200, Orri Erling wrote:

  

Has has there been any analysis on whether there is a *fundamental*
reason for such performance difference? Or is it simply a question of
maturity; in other words, relational db technology has been around
  

for a
  

very long time and is very mature, whereas RDF implementations are
  

still
  

quite recent, so this gap will surely narrow ...?

  

This is a very complex subject.  I will offer some analysis below, but
this I fear will only raise further questions.  This is not the end of


the
  

road, far from it.
  


As far as I understand, another issue is relevant: this benchmark is
somewhat unfair as the relational stores have one advantage compared to
  

the
  

native triple stores: the relational data structure is fixed (Products,
Producers, Reviews, etc with given columns), while the triple
  

representation
  

is generic (arbitrary s,p,o).

  

This point has an effect on several levels.

For instance, the flexibility afforded by triples means that objects
stored in this structure require processing just to piece it all
together, whereas the RDBMS has already encoded the structure into the
table. Ironically, this is exactly the reason we
(Tucana/Kowari/Mulgara) ended up building an RDF database instead of
building on top of an RDBMS: The flexibility in table structure was
less efficient that a system that just knew it only had to deal with
3 columns. Obviously the shape of the data (among other things)
dictates what it is the better type of storage to use.

A related point is that processing RDF to create an object means you
have to move around a lot in the graph. This could mean a lot of
seeking on disk, while an RDBMS will usually find the entire object in
one place on the disk. And seeks kill performance.

This leads to the operations used to build objects from an RDF store.
A single object often requires the traversal of several statements,
where the object of one statement becomes the subject of the next.
Since the tables are typically represented as
Subject/Predicate/Object, this means that the main table will be
joined against itself. Even RDBMSs are notorious for not doing this
efficiently.

One of the problems with self-joins is that efficient operations like
merge-joins (when they can be identified) will still result in lots of
seeking, since simple iteration on both sides of the join means
seeking around in the same data. Of course, there ARE ways to optimize
some of this, but the various stores are only just starting to get to
these optimizations now.

Relational databases suffer similar problems, but joins are usually
only required for complex

Re: Announcing Open GUID

2008-09-25 Thread Kingsley Idehen


[EMAIL PROTECTED] wrote:

Yes, I have not loaded DBPedia yet. I started with WordNet 3.0, which
apparently did not deem IBM as a part of our language (note Mircosoft
and Google do make the cut!). See the (http://openguid.net/roadmap) for
more details on the data loading initiative. Once that is complete, if
your favorite (least favorite?) company is still not loaded, anyone will
be able to add it, tag it, and add identical references.

And yes, the function was to find you an existing Open GUID based on
some text. I will try to make that clearer.

Thanks for the test drive,
Jason
  

Jason,

Yes, I see some data now [1].
How are you planning to proceed re. data from UMBEL, Yago, OpenCyc, and 
DBpedia?



Links:

1. 
http://demo.openlinksw.com/proxy/html/http://openguid.net/e6adf4e8-da25-102b-9a03-2db401e887ec



Kingsley

 Original Message 
Subject: Re: Announcing Open GUID
From: Kingsley Idehen [EMAIL PROTECTED]
Date: Thu, September 25, 2008 8:48 am
To: Damian Steer [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED], public-lod@w3.org


Damian Steer wrote:
  

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Kingsley Idehen wrote:

| I am exploring your service.
|
| Please clarify this:
|
http://openguid.net/search?keyword=http%3A%2F%2Fdbpedia.org%2Fresource%2FIBM 



|
|
| The demo asked me to input an Open GUID (a URI) re. the Find feature.

Do you mean 'Find GUID'? I read that as meaning 'Find me a GUID for this
text I've just entered'.


Damian,

Yes.

Okay, I tried: http://openguid.net/search?keyword=ibm

Didn't get a URI back.

Sindice give me: http://sindice.com/search?q=ibmqt=term

Maybe the DBpedia binding hasn't take place yet? I am hoping to get the 
URIs associated with IBM the entity of Type 
some-ontology:some-term-example-organization


Kingsley
  

Damian
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.6 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFI25r3AyLCB+mTtykRAm50AKCMvpRUjv0XOabY7Q9Kcg9kuAV/7gCfZj2t
o1plwBi8zNimV+UGHmxAL2k=
=G1Ix
-END PGP SIGNATURE-





  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Drilling into the LOD Cloud

2008-09-28 Thread Kingsley Idehen


Dan Brickley wrote:

Kingsley Idehen wrote:


Ed Summers wrote:

On Mon, Sep 22, 2008 at 1:28 PM, Giovanni Tummarello
[EMAIL PROTECTED] wrote:
 

I really support your idea for a lod cloud that's actually useful to
write queries, i promise we'll do the best from sindice to deliver one
such a thing.



That sounds like a great idea for a service. Right now I just need a
little diagram that illustrates the links between resources in the LOD
cloud; and the diversity of descriptions from each provider. But, I
guess I should stop begging on here, and just create it eh?

Thanks for the feedback,
//Ed


  

Ed,

Are you not able to add additional rdf:type links between DBpedia 
and the following:


1. Yago
2. OpenCyc
3. UMBEL

Then between UMBEL and OpenCyc:

1. owl:sameAs
2. owl:equivalentClass


If these thingies are owl:sameAs, then presumably they have same 
IP-related characteristics, owners, creation dates etc?


Does that mean Cycorp owns UMBEL?

Dan,

No, it implies that in the UMBEL data space you have equivalence between 
Classes used to define UMBEL subject concepts (subject matter entities) 
and OpenCyc.


Simple examples via SPARQL against DBpedia using the SPARQL endpoint: 
http://dbpedia.org/sparql :


-- Get DBpedia Entities of Type opencyc:Motorcycle where the 
dbpedia:name contains pattern: BMW


define input:inference 'http://dbpedia.org/resource/inference/rules/umbel#'
prefix umbel: http://umbel.org/umbel/sc/
prefix dbpedia: http://dbpedia.org/property/
prefix opencyc: http://sw.opencyc.org/2008/06/10/concept/en/
select ?s
where
{
?s a opencyc:Motorcycle.
?s dbpedia:name ?n.
?n bif:contains BMW.
}



-- Get DBpedia Entities of Type umbel:Motorcycle where the dbpedia:name 
contains pattern: BMW


define input:inference 'http://dbpedia.org/resource/inference/rules/umbel#'
prefix umbel: http://umbel.org/umbel/sc/
prefix dbpedia: http://dbpedia.org/property/
prefix opencyc: http://sw.opencyc.org/2008/06/10/concept/en/
select ?s
where
{
?s a umbel:Motorcycle.
?s dbpedia:name ?n.
?n bif:contains BMW.
}


Both queries return the same results even though in DBpedia you only 
have entities of type umbel:Motorcycle as shown here:

http://dbpedia.org/resource/BMW_R75 .

Courtesy of UMBEL and UMBEL inference rules applied to DBpedia, we can 
now expand the scope of existing DBpedia entities by exploiting 
relationships that exist across the Classes exposed via rdf:type. 
Thus, via rdf:type we are able to cross over from the instance realm 
to the data dictionary realm, and once in the data dictionary realm 
expand our horizons  via  class level relationships  covering 
equivalence and the broader and narrower transitive relationships in the 
Class Hierarchy.


So OpenCyc doesn't own UMBEL, it is simply associated with UMBEL by 
being part of the data dictionary oriented Linked Data Space that UMBEL 
delivers :-)


Kingsley



Then between OpenCyc and Wordnet:
1. owl:sameAs


Ditto...

cheers,

Dan




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Role of RDF on the Web and within enterprise applications. was: AW: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL

2008-09-30 Thread Kingsley Idehen
the views, thus a high penalty for wrong guess.  Andif it is hard enough
to figure out where a query plan goes wrong with a given schema, it is
harder still to figure it out with a schema that morphs by itself.

In the RDB world, for example Oracle recommends saving optimizer statistics
from the  test  environment and using these in the production environment
just so the optimizer does not get creative.  Now this is the  essence of
wisdom for OLTP but we are not talking OLTP with RDF. 


If there is a history of usage and this history is steady and the dba can
confirm it as being a representative sample, then automatic materializing
of joins is a real  possibility.  Doing this spontaneously would lead to
erratic response times, though.  For anything online, the accent is more on
predictable throughput than peak throughput.  


The BSBM query mix does lend itself quite well to automatic materialization
but with this, one would not normally do the representation as RDF to begin
with, the workload being so typically relational.  Publishing any ecommerce
database as RDF is of course good but then mapping is the simpler and more
predictable route.  


It is my feeling that RDF has a dual role:  1. interchange format:  This is
like what XML does, except that RDF has more semantics and expressivity.  2:
Database storage format for cases where data must be integrated and is too
heterogenous to easily fall into one relational schema.  This is for example
the case in the open web conversation and social space.  The first case is
for mapping, the second for warehousing.   Aside this, there is potential
for more expressive queries through  the query language dealing with
inferencing, like  subclass/subproperty/transitive etc.  These do not go
very well with SQL views.



Orri












  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Linked Data Oriented Submissions to the Billion Triples Challenge

2008-10-01 Thread Kingsley Idehen


All,

So far,  from the LOD camp, we have:

1. 
http://www.openlinksw.com/dataspace/oerling/weblog/Orri%20Erling%27s%20Blog/1445

2.  http://www.freebase.com/view/user/bio2rdf/public/sparql

Stats re. Bio2Rdf.org submission:

Here are the final numbers


http://www.freebase.com/view/user/bio2rdf/public/sparql

2,345,752,436 or 2.3 Billions triples

available from those 27 sparql endpoints

http://biocyc.bio2rdf.org/sparql
http://biopax.bio2rdf.org/sparql
http://chebi.bio2rdf.org/sparql
http://cpath.bio2rdf.org/sparql
http://dbpedia.bio2rdf.org/sparql
http://ec.bio2rdf.org/sparql
http://geneid.bio2rdf.org/sparql
http://go.bio2rdf.org/sparql
http://hgnc.bio2rdf.org/sparql
http://homologene.bio2rdf.org/sparql
http://inoh.bio2rdf.org/sparql
http://iproclass.bio2rdf.org/sparql
http://kegg.bio2rdf.org/sparql
http://mesh.bio2rdf.org/sparql
http://mgi.bio2rdf.org/sparql
http://obo.bio2rdf.org/sparql
http://omim.bio2rdf.org/sparql
http://pdb.bio2rdf.org/sparql
http://protein.bio2rdf.org/sparql
http://pubchem.bio2rdf.org/sparql
http://pubmed.bio2rdf.org/sparql
http://reactome.bio2rdf.org/sparql
http://taxonomy.bio2rdf.org/sparql
http://uniparc.bio2rdf.org/sparql
http://uniprot.bio2rdf.org/sparql
http://uniref.bio2rdf.org/sparql
http://unists.bio2rdf.org/sparql

it is 38 public datasources that are globally made available in RDF via  
Virtuoso SPARQ endpoint.


uniref, uniparc and pubmed will be uploaded during the next weeks





--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Watchdog.net is plugged into the DBpedia-LOD Cloud

2008-10-15 Thread Kingsley Idehen


Anja,

http://watchdog.net should be added to the new DBpedia-Cloud graphic. It 
gets to DBpedia via GovTrack linkage.


See: 
http://linkeddata.uriburner.com/about/html/http://watchdog.net/p/nancy_pelosi
Note: The value of predicate : 
http://watchdog.net/about/api#govtrackid is now a GovTrack URI as 
opposed to the initial literal value.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Watchdog.net is plugged into the DBpedia-LOD Cloud

2008-10-15 Thread Kingsley Idehen


Dan Brickley wrote:


Aaron Swartz wrote:
http://www.rdfabout.com/rdf/usgov/congress/people/P000197 which is 
in my


Those are bioguideids, right?

Kingsley seems to have some sort of rewrite that converts from
govtrack.us URLs to the right thing and since he's the only user, I
decided to do it that way.


I've lost track of who is behind what here, but looking at above URL I 
see


 xmlns:ns=http://xmlns.com/foaf/0.1/;

...

Besides FOAF Person, gender, img, which are all fine, I see a 
foaf:religion unknown or proposed property:


ns:religionRoman Catholic/ns:religion

...this isn't in the spec, but I'd consider it for sure if it's 
proposed on foaf-dev.



Also an rdf syntax issue:
ns:homepagehttp://www.house.gov/pelosi/ns:homepage
should be...
ns:homepage rdf:resource=http://www.house.gov/pelosi/

cheers,

Dan








Dan,

Aaron is behind Watchdog.net .

Joshua is behind GovTrack.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Watchdog.net is plugged into the DBpedia-LOD Cloud

2008-10-15 Thread Kingsley Idehen


Joshua Tauberer wrote:

Kingsley Idehen wrote:
http://watchdog.net should be added to the new DBpedia-Cloud graphic. 
It gets to DBpedia via GovTrack linkage.


See: 
http://linkeddata.uriburner.com/about/html/http://watchdog.net/p/nancy_pelosi 

Note: The value of predicate : 
http://watchdog.net/about/api#govtrackid is now a GovTrack URI as 
opposed to the initial literal value.


Actually the URI isn't quite right. (I can hear Aaron and Kingsley 
going d'oh after all of the (off-list) emails about this.)


Aaron- I suggest that #govtrackid go back to its original value (e.g. 
400314) and add a owl:sameAs to 
http://www.rdfabout.com/rdf/usgov/congress/people/P000197 which is 
in my URI space for politicians, as opposed to 
http://www.govtrack.us/congress/person.xpd?id=400314 which is simply 
the HTML page on GovTrack.



Joshua,

Why not have both links (one for the page association and owl:sameAs for 
the Person Entity links) ? The document link is but one association 
across the data spaces that you an Aaron maintain :-) maybe #govtrackid 
could become #govtrack_page leaving the URI value unchanged.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: web to semantic web : an automated approach

2008-10-26 Thread Kingsley Idehen


रविंदर ठाकुर (ravinder thakur) wrote:

Hello Friends,


I have been following semantic web for some time now and have seen 
quite a lot of projects being run (dbpedia, FOAF, LOD etc) trying to 
generate/organize some semantic content. While these approaches might 
have been successful in their goals, one major problem plaguing 
semantic web as a whole is the lack of semantic content. Unfortunately 
there is nothing in sight that we can rely on to generate semantic 
content for the truckloads of information being put on web everyday. I 
strongly feel that one of the _wrong_ assumption in semantic web 
community is that content creators will be creating a semantic data. 
This I think is too much for the asking from even more technically 
sound part of web community let along whole of the web community. It 
hasn't happened over last so many years and I don't see it happening 
in the near future.



To really move the semantic web forward is a mechanism to device a 
mechanism to _automatcially_ convert the information over the web to 
semantic information. There are many softwares/services that can be 
used for this purpose. I am currently developing one prototype for 
this purpose. This prototype uses services from 
OpenCalais(http://www.opencalais.com/) to convert ordinary text to 
semantic form. This service is very limited in what entities supports 
at the moment but its a very good start. I am pretty sure there will 
be many other good options available that might be unknown to me. The 
currently very primitive prototype can be seen at 
http://arcse.appspot.com. This currently implements very few of the 
ideas I have for this. This is hosted on Google's AppEngine so 
sometime gives timeout messages internally so please bear with this :).



This automatic conversion however is not a simple task and needs work 
in lot in domains ranging form NLP to artificial intelligence to 
semantic web to logic etc. So that's why this mail. I will be more 
than happy if we can join together to form a like minded team that can 
work on solving this most important problem plaguing semantic web 
currently.



Waiting for your suggestions/criticisms. And Happy Diwali 
http://en.wikipedia.org/wiki/Diwali too :)

Ravinder Thakur


PS : I posted similar query here 
http://www.nabble.com/web-to-semantic-web-%3A-an-automated-approach-td20064203.htmlas 
well. That generated some good debate.

Ravinder,

Please read the various posts I've made about RDF Middleware which is 
all about the recently branded top-down approach to the Linked Data 
Web. From day one, I've always subscribed to the belief that structured 
linked data will need to be generated from existing Web content in order 
for the Linked Data aspect of the Web to blossom.


Links:

1. 
http://www.openlinksw.com/dataspace/[EMAIL PROTECTED]/weblog/[EMAIL PROTECTED]/1454
2. 
http://www.openlinksw.com/weblog/public/search.vspx?blogid=127q=rdf%20middlewaretype=textoutput=html

3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtSponger

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Freebase Linked Data

2008-10-31 Thread Kingsley Idehen


All,

A few samplings of what the Freebase Linked Data space offers courtesy 
of our Linked Data Explorer offerings (Firefox extension [1] or Live 
hosted service [2]) .


1. http://tinyurl.com/5mhnsu -- About Obama
2. http://tinyurl.com/6br773 -- About Hillary Clinton

Note: We connect the container (information resource) with the relevant 
entities via foaf:primarytopic. Also note the use of UMBEL (but much 
more to come on this later now that ground zero of the Linked Data has 
more or less crystallized).


1. http://ode.openlinksw.com
2. http://linkeddata.uriburner.com/ode (*what used to be 
http://demo.openlinksw.com/rdfbrowser2 | rdfbrowser *)



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Linked Data Tutorial Material

2008-11-14 Thread Kingsley Idehen


Ed Summers wrote:

On Fri, Nov 14, 2008 at 6:40 AM, Yves Raimond [EMAIL PROTECTED] wrote:
  

We plan to publish, with Keith, a small how-to for doing a hands-on
tutorial like the Web-of-data 101 session we did at the WOD-PD event.
With Richard's permission, we'll also take a few things from his
session, where he used some of the data we created during our session.

The goal of this how-to is to make it easy for people to reproduce
the tutorial in different places.



This would be extremely helpful to me and some others who are thinking
of doing a linked-data tutorial at code4lib2009 [1]. Please post a
message here when any of the materials materialize!

//Ed

[1] http://code4lib.org/2009


  

All,

Shouldn't we be able to use a Wiki to produce and maintain a series of 
Linked Data Tutorials aimed at a plethora of target audiences?


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen
-sc:Graduate
   * umbel-sc:Communicator
   * umbel-sc:Person
   * umbel-sc:Entertainer


Links:

1. http://sw.opencyc.com
2. http://umbel.org


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Dbpedia-announcements] ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen


Chris Bizer wrote:

Hi Andreas,

we for sure want to do this, but also did not want to postpone the DBpedia
3.2 release any further.

So be ensured that the upcoming public user interface for defining the
infobox-to-ontology mappings will include the possibility to reuse existing
classes and properties and that external classes and properties will be used
within the 3.3 release.

Defining the infobox-to-ontology mappings that we currently have was already
a lot of work (Anja thanks again), so please be patient with the
mappings/reuse of external ontologies.

Cheers

Chris
  

Andreas,

Apropos Chris' comments above, it's coming.

DBpedia ontology to UMBEL ontology mapping is one of a number of 
mappings that will emerge in due course. That said -- by the very nature 
of this project and RDF based Linked Data in general -- anyone with a 
vested interest in the ontology mapping efforts can step in and 
accelerate matters. Thus, I encourage others to participate bearing in 
mind DBpedia 3.2's increased compatibility with such endeavors.


Kingsley
 

  

-Ursprüngliche Nachricht-
Von: Andreas Harth [mailto:[EMAIL PROTECTED]
Gesendet: Montag, 17. November 2008 16:55
An: Chris Bizer
Cc: public-lod@w3.org; 'Semantic Web'; dbpedia-
[EMAIL PROTECTED]; dbpedia-
[EMAIL PROTECTED]
Betreff: Re: ANN: DBpedia 3.2 release, including DBpedia Ontology and
RDF links to Freebase

Hi Chris,

Chris Bizer wrote:


1. DBpedia Ontology

DBpedia now features a shallow, cross-domain ontology, which has been
manually created based on the most commonly used infoboxes within
  

Wikipedia
great work!

One thing: what's the reason for creating your own classes rather
than re-using or sub-classing existing ones (foaf:Person,
geonames:Feature...)?  Same for properties (foaf:name, dc:date...).

Regards,
Andreas.

--
http://swse.deri.org/




-
This SF.Net email is sponsored by the Moblin Your Move Developer's challenge
Build the coolest Linux based applications with Moblin SDK  win great prizes
Grand prize is a trip for two to an Open Source event anywhere in the world
http://moblin-contest.org/redirect.php?banner_id=100url=/
___
Dbpedia-announcements mailing list
[EMAIL PROTECTED]
https://lists.sourceforge.net/lists/listinfo/dbpedia-announcements

  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen


John Goodwin wrote:
  

Have fun with the new DBpedia knowledge base!

Cheers

Chris



Thanks Chris and team for all your hard work getting this done. I do,
however, have a few comments regarding the OWL ontology. I think in
general the use of domain and range is perhaps a bit dubious in that
for many things I think it is overly specified. I can imagine anyone
re-using the Dbpedia properties getting some unexpected inferences from
the domain and range restrictions. Also the range restriction seem to be
done as an OWL intersection so if, for example, something has a
publisher x then x will be inferred to be both a Company and a Person
which is probably not what you want. Personally, in all but a few cases,
I'd be tempted to generalise or just remove the domain/range
restrictions. Any thoughts?

john
.

  

John,

Yes, this issues exist with the DBpedia Ontology, but also bear in mind 
this is also why UMBEL and DBpedia have been Linked :-)


Eventually (as per my other mail posts), the DBpedia Ontology will need 
to be linked with UMBEL (*which isn't the case right now*) so that it 
can benefit from the work done in the UMBEL project. That said, UMBEL is 
also a loosely bound ontology for DBpedia already, so you can also 
optionally leverage for inference purposes  right now.  All in all, I am 
hoping a lot of the inherent flexibilities that the Internet hosted 
Distributed Database we know as the Web accords start to come through 
with much more clarity following DBpedia 3.2.


Kingsley

This email is only intended for the person to whom it is addressed and may 
contain confidential information. If you have received this email in error, 
please notify the sender and delete this email which must not be copied, 
distributed or disclosed to any other person.

Unless stated otherwise, the contents of this email are personal to the writer 
and do not represent the official view of Ordnance Survey. Nor can any contract 
be formed on Ordnance Survey's behalf via email. We reserve the right to 
monitor emails and attachments without prior notice.

Thank you for your cooperation.

Ordnance Survey
Romsey Road
Southampton SO16 4GU
Tel: 08456 050505
http://www.ordnancesurvey.co.uk



  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Dbpedia-discussion] ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen
 the information in any way nor act in reliance on 
it and notify the sender immediately.
Please note that the BBC monitors e-mails sent or received.
Further communication will signify your consent to this.


-
This SF.Net email is sponsored by the Moblin Your Move Developer's challenge
Build the coolest Linux based applications with Moblin SDK  win great prizes
Grand prize is a trip for two to an Open Source event anywhere in the world
http://moblin-contest.org/redirect.php?banner_id=100url=/
___
Dbpedia-discussion mailing list
[EMAIL PROTECTED]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen


Jens Lehmann wrote:

Hello John,

John Goodwin wrote:
  

Thanks Chris and team for all your hard work getting this done. I do,
however, have a few comments regarding the OWL ontology. I think in
general the use of domain and range is perhaps a bit dubious in that
for many things I think it is overly specified. I can imagine anyone
re-using the Dbpedia properties getting some unexpected inferences from
the domain and range restrictions. Also the range restriction seem to be
done as an OWL intersection so if, for example, something has a
publisher x then x will be inferred to be both a Company and a Person
which is probably not what you want. Personally, in all but a few cases,
I'd be tempted to generalise or just remove the domain/range
restrictions. Any thoughts?



We specified the domains and ranges as disjunctions of classes (not
intersection). See the W3C specification of owl:unionOf [1].

The domain and range axioms help to structure DBpedia and clarify the
meaning of certain properties. While there is room for improvement, it
is not an option to remove all of them.

Currently, there are two versions of the infobox extraction: a loose one
and a strict one. In the strict one, it is guaranteed that the data
complies to the ranges specified in the ontology schema. Currently, only
the loose (probably inconsistent) one is provided.

Kind regards,

Jens

[1] http://www.w3.org/TR/owl-guide/#owl_unionOf


  

Jens,

What's the URL of the strict one?

We are building a DBpedia installer for Virtuoso, so at the very least I 
want the users of this installer to have choice of strict or loose 
infobox extraction.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen


Georgi Kobilarov wrote:

Kingsley,
 
  

What's the URL of the strict one?

We are building a DBpedia installer for Virtuoso, so at the very least
I
want the users of this installer to have choice of strict or loose
infobox extraction.



Not publicly available yet. There was a buggy first version of strict,
but we decided to no further work on it for release 3.2. 

  

Okay.

Kingsley

Georgi

--
Georgi Kobilarov
Freie Universität Berlin
www.georgikobilarov.com


  

-Original Message-
From: Kingsley Idehen [mailto:[EMAIL PROTECTED]
Sent: Monday, November 17, 2008 6:14 PM
To: Jens Lehmann
Cc: public-lod@w3.org; Semantic Web; dbpedia-
[EMAIL PROTECTED]; John Goodwin
Subject: Re: [Dbpedia-discussion] DBpedia 3.2 release, including
DBpedia Ontology and RDF links to Freebase

Jens Lehmann wrote:


Hello John,

John Goodwin wrote:

  

Thanks Chris and team for all your hard work getting this done. I


do,


however, have a few comments regarding the OWL ontology. I think in
general the use of domain and range is perhaps a bit dubious in


that


for many things I think it is overly specified. I can imagine


anyone
  

re-using the Dbpedia properties getting some unexpected inferences


from


the domain and range restrictions. Also the range restriction seem


to be


done as an OWL intersection so if, for example, something has a
publisher x then x will be inferred to be both a Company and a


Person


which is probably not what you want. Personally, in all but a few


cases,


I'd be tempted to generalise or just remove the domain/range
restrictions. Any thoughts?



We specified the domains and ranges as disjunctions of classes (not
intersection). See the W3C specification of owl:unionOf [1].

The domain and range axioms help to structure DBpedia and clarify
  

the
  

meaning of certain properties. While there is room for improvement,
  

it


is not an option to remove all of them.

Currently, there are two versions of the infobox extraction: a loose
  

one


and a strict one. In the strict one, it is guaranteed that the data
complies to the ranges specified in the ontology schema. Currently,
  

only


the loose (probably inconsistent) one is provided.

Kind regards,

Jens

[1] http://www.w3.org/TR/owl-guide/#owl_unionOf



  

Jens,

What's the URL of the strict one?

We are building a DBpedia installer for Virtuoso, so at the very least
I
want the users of this installer to have choice of strict or loose
infobox extraction.

--


Regards,

Kingsley Idehen   Weblog:


http://www.openlinksw.com/blog/~kidehen
  

President  CEO
OpenLink Software Web: http://www.openlinksw.com








---
  

--
This SF.Net email is sponsored by the Moblin Your Move Developer's
challenge
Build the coolest Linux based applications with Moblin SDK  win great
prizes
Grand prize is a trip for two to an Open Source event anywhere in the
world
http://moblin-contest.org/redirect.php?banner_id=100url=/
___
Dbpedia-discussion mailing list
[EMAIL PROTECTED]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion



  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-17 Thread Kingsley Idehen
 about a thing from DBpedia and Freebase.

For more information about the Freebase links see:
http://blog.dbpedia.org/2008/11/15/dbpedia-is-now-interlinked-with-
freebase-
links-to-opencyc-updated/


3. Cleaner Abstacts

Within the old DBpedia dataset it occurred that the abstracts for
different
languages contained Wikpedia markup and other strange characters. For
the
3.2 release, we have improved DBpedia's abstract extraction code which
results in much cleaner abstracts that can safely be displayed in user
interfaces.


The new DBpedia release can be downloaded from:

http://wiki.dbpedia.org/Downloads32

and is also available via the DBpedia SPARQL endpoint at

http://dbpedia.org/sparql

and via DBpedia's Linked Data interface. Example URIs:

http://dbpedia.org/resource/Berlin
http://dbpedia.org/page/Oliver_Stone

More information about DBpedia in general is found at:

http://wiki.dbpedia.org/About


Lots of thanks to everybody who contributed to the Dbpedia 3.2


release!
  

Especially:

1. Georgi Kobilarov (Freie Universität Berlin) who designed and
implemented
the new infobox extraction framework.
2. Anja Jentsch (Freie Universität Berlin) who contributed to
implementing
the new extraction framework and wrote the infobox to ontology class
mappings.
3. Paul Kreis (Freie Universität Berlin) who improved the datatype
extraction code.
4. Andreas Schultz (Freie Universität Berlin) for generating the
Freebase to
DBpedia RDF links.
5. Everybody at OpenLink Software for hosting DBpedia on a Virtuoso
server
and for providing the statistics about the new Dbpedia knowledge base.

Have fun with the new DBpedia knowledge base!

Cheers

Chris


--
Prof. Dr. Christian Bizer
Web-based Systems Group
Freie Universität Berlin
+49 30 838 55509
http://www.bizer.de
[EMAIL PROTECTED]






-
This SF.Net email is sponsored by the Moblin Your Move Developer's challenge
Build the coolest Linux based applications with Moblin SDK  win great prizes
Grand prize is a trip for two to an Open Source event anywhere in the world
http://moblin-contest.org/redirect.php?banner_id=100url=/
___
Dbpedia-discussion mailing list
[EMAIL PROTECTED]
https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion

  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: AW: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase

2008-11-18 Thread Kingsley Idehen
 that are present in a certain
dataset. They simply express what's there in the data. In this sense,
they are like database schemas: If Publisher has a range of
Person, then it means that the publisher *in this particular
dataset* is always a person. That's not an assertion about the world,
it's an assertion about the dataset. These ontologies are usually not
very re-usable.

b) Ontologies that are intended as a lingua franca for data exchange
between different applications. They are designed for broad re-use,
and thus usually do not add many restrictions. In this sense, they are
more like controlled vocabularies of terms. Dublin Core is probably
the prototypical example, and FOAF is another good one. They usually
don't allow as many interesting inferences.

I think that these two kinds of ontologies have very different
requirements. Ontologies that are designed for one of these roles are
quite useless if used for the other job. Ontologies that have not been
designed for either of these two roles usually fail at both.

Returning to DBpedia, my impression is that the DBpedia ontology is
intended mostly for the first role. Maybe it should be understood more
as a schema for the DBpedia dataset, and not so much as a re-usable
set of terms for use outside of the Wikipedia context. (I might be
wrong, I was not involved in its creation.)

Richard




  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com







Re: Suggestions/existing material for Linked Data tutorial?

2008-11-19 Thread Kingsley Idehen


Yves Raimond wrote:

Hello!

  

We plan to publish, with Keith, a small how-to for doing a hands-on
tutorial like the Web-of-data 101 session we did at the WOD-PD event.
With Richard's permission, we'll also take a few things from his
session, where he used some of the data we created during our session.

The goal of this how-to is to make it easy for people to reproduce
the tutorial in different places.




I (finally) started this linked data tutorial how-to, at:
http://community.linkeddata.org/MediaWiki/index.php?WODPDTutorial

Still a lot to do though - I didn't even touch the second part of the
tutorial. I also wrote a small script, linked from this Wiki page, and
implementing a small linked data server [1] - attendees can POST RDF
data to it, and it gets published as linked data. It also provides a
SPARQL end-point. Anyway, early comments are welcome :-)

I hope to finish that soon (I need to do this tutorial at the BBC
soonish, so it'll have to be ready by then :-) )

Cheers!
y

[1] http://moustaki.org/resources/rdf-pastebin.tar.gz


  

Yves,

Nice start.

We will add our own tutorial which starts from: Get Yourself a URI in 5 
mins  of less :-)




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Size matters -- How big is the danged thing

2008-11-21 Thread Kingsley Idehen


David Wood wrote:
Sorry to intervene here, but I think Kingsley's suggestion sets up a 
false dicotomy. REST principles (surely part of everything we stand 
for :) suggest that the source of RDF doesn't matter as long as a URL 
returns what we want. Late binding means not having to say you're sorry.


Is it a good idea to set up a class system where those who publish to 
files are somehow better (or even different!) than those who publish 
via adapters?

David,

Yes, the dichotomy is false if the basis is: Linked Data irrespective of 
means or source, as long as the URIs are de-referencable. On the other 
hand, if Linked Data generated on the fly isn't deemed part of the LOD 
cloud (the qualm expressed in Giovanni's comments) then we have to call 
RDF-ized Linked Data something :-)


You can count the warehouse (an arrive at hub size) but the RDF-ized 
stuff is a complete red herring (imho - see cool fractal animations post).


What I am hoping is a more interesting quesion is this: have we reached 
the point were we can drop burgeoning from the state of the Linked 
Data Web? Do we have a hub that provides enough critical mass for the 
real fun to start (i.e., finding stuff with precision that data object 
properties accord) ?


Personally, I think the Linked Data Web has reached this point, so our 
attention really has to move more towards showing what Linked Data adds 
to the Web in general.



Kingsley



So, I vote for counting all of it. Isn't that what Google and Yahoo do 
when they count the number of pages indexed?


Regards,
Dave
--

On Nov 21, 2008, at 4:26 PM, Kingsley Idehen [EMAIL PROTECTED] 
wrote:




Giovanni Tummarello wrote:

Overall, that's about 17 billion.




IMO considering myspace 12 billion triples as part of LOD, is quite a
stretch (same with other wrappers) unless they are provided by the
entity itself (E.g. i WOULD count in livejournal foaf file on the
other hand, ok they're not linked but they're not less useful than the
myspace wrapper are they? (in fact they are linked quite well if you
use the google social API)


Giovanni




Giovanni,

Maybe we should use the following dichotomy re. the Web of Linked 
Data (aka. Linked Data Web):


1. Static Linked Data or Linked Data Warehouses - which is really 
what the LOD corpus is about
2. Dynamic Linked Data - which is what RDF-zation middleware 
(including wrapper/proxy URI generators) is about.


Thus, I would say that Jim is currently seeking stats for the Linked 
Data Warehouse part of the burgeoning Linked Data Web. And hopefully, 
once we have the stats, we can get on to the more important task of 
explaining and demonstrating the utility of the humongous Linked Data 
corpus :-)


ESW Wiki should be evolving as I write this mail (i.e. tabulated 
presentation of the data that's already in place re. this matter).



All: Could we please stop .png and .pdf based dispatches of data, it 
kinda contradicts everything we stand for :-)


--


Regards,

Kingsley Idehen  Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO OpenLink Software Web: http://www.openlinksw.com










--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Size matters -- How big is the danged thing

2008-11-21 Thread Kingsley Idehen


Aldo Bucchi wrote:

On Fri, Nov 21, 2008 at 7:51 PM, Kingsley Idehen [EMAIL PROTECTED] wrote:
  

Yves Raimond wrote:


On Fri, Nov 21, 2008 at 8:08 PM, Giovanni Tummarello
[EMAIL PROTECTED] wrote:

  

Overall, that's about 17 billion.


  

IMO considering myspace 12 billion triples as part of LOD, is quite a
stretch (same with other wrappers) unless they are provided by the
entity itself (E.g. i WOULD count in livejournal foaf file on the
other hand, ok they're not linked but they're not less useful than the
myspace wrapper are they? (in fact they are linked quite well if you
use the google social API)



Actually, I don't think I can agree with that. Whether we want it or
not, most of the data we publish (all of it, apart from specific cases
e.g. review) is provided by wrappers of some sort, e.g. Virtuoso, D2R,
P2R, web services wrapper etc. Hence, it makes not sense trying to
distinguish datasets on the basis they're published through a
wrapper or not.

Within LOD, we only segregate datasets for inclusion in the diagram on
the basis they are published according to linked data principles. The
stats I sent reflect just that: some stats about the datasets
currently in the diagram.

The origin of the data shouldn't matter. The fact that it is published
according to linked data principles and linked to at least one dataset
in the cloud should matter.



  

Giovanni





  

Yves,

I agree. But I am sure you can also see the inherent futility in pursuing
the size of the pure Linked Data Web :-)  The moment you arrive at a number
it will be obsolete :-)

I would frame the question this way: is LOD hub now dense enough for basic
demonstrations of Linked Data Web utility to everyday Web users? For
example, can we Find stuff on the Web with levels of precision and
serendipity erstwhile unattainable? Can we now tag stuff on the Web in a
manner that makes tagging useful? Can we alleviate the daily costs of Spam
on mail inboxes? Can all of the aforementioned provide the basis for
relevant discourse discovery and participation?



Sorry, this is getting too interesting to stay in lurker mode ;)

Kingsley, absolutely. We have got to that point. The fun part has begun.

To quote Jim, who started this thread:

http://blogs.talis.com/nodalities/2008/03/jim_hendler_talks_about_the_se.php

Go to minute 28 aprox ( I can't listen to it here, I just blocked mp3's ).
Jim touches on how a geo corpus can be used to dissambiguate tags on flickr.
This is one such use, low hanging fruit wrt the huge amount of linked
data, and a first timer in terms of IT.

This was not possible last year!
It is now.

I guess that is THE question now: What can we do this year that we
couldn't do last year?
( thanks to the massive amount of available LOD ).

Best,
A
  

Aldo,

Yep!

So we should start building up a simple collection (in a Wiki) of simple 
and valuable things you can now achieve courtesy of Linked Data :-)


Find replacing Search as the apex of the Web value proposition 
pyramid for everyday Web Users.


Courtesy of Linked Data (warehouse and/or dynamic), every Web 
information resource is now a DBMS View in disguise :-)


Kingsley
  

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO OpenLink Software Web: http://www.openlinksw.com











  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Size matters -- How big is the danged thing

2008-11-21 Thread Kingsley Idehen


David Wood wrote:


On Nov 21, 2008, at 5:51 PM, Kingsley Idehen wrote:
I would frame the question this way: is LOD hub now dense enough for 
basic demonstrations of Linked Data Web utility to everyday Web 
users? For example, can we Find stuff on the Web with levels of 
precision and serendipity erstwhile unattainable? Can we now tag 
stuff on the Web in a manner that makes tagging useful? Can we 
alleviate the daily costs of Spam on mail inboxes? Can all of the 
aforementioned provide the basis for relevant discourse discovery and 
participation?


An interesting experiment might be to start at some bit of RDF (a FOAF 
document or some such) and follow-your-nose from link to see to see 
how far the longest path is.  If it is very, very long (maybe even 
nicely loopy since the LOD effort), then life is good.


Regards,
Dave







Dave,

That's what this is all about:
http://b3s.openlinksw.com/  ( a huge Linked Data corpus, talking 11 
Billion or so triples).  What was missing from this demo all along was a 
Find  feature that hides all the SPARQL :-)


Also, there will be more when we finally release the long overdue update 
to the OpenLink Data Explorer :-)



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: linked data mashups

2008-11-23 Thread Kingsley Idehen


Peter Ansell wrote:
2008/11/23 Juan Sequeda [EMAIL PROTECTED] 
mailto:[EMAIL PROTECTED]


Hi Giovanni and all


On Sat, Nov 22, 2008 at 7:33 PM, Giovanni Tummarello
[EMAIL PROTECTED]
mailto:[EMAIL PROTECTED] wrote:


 I guess that is THE question now: What can we do this year
that we
 couldn't do last year?
 ( thanks to the massive amount of available LOD ).

Two days ago the discussion touched this interesting point. I
do not
know how to answer this question.
Ideas?


We need to start consuming linked data and making reall mashup
applications powered by linked data. A couple of days I just
mentioned the link for SQUIN: http://squin.sourceforge.net/

The idea of SQUIN came out of ISWC08 with Olaf Hartig. The
objective is to make LOD accesible easily to web2.0 app
developers. We envision adding an S compoment to the LAMP stack.
This will allow people to easily query LOD from their own server.

We should have a demo ready in the next couple of weeks.

We believe that this is something needed to actually start using
LOD and making it accesible to everybody.


How does SQIUN differ to a typical HTTP SPARQL endpoint? So far it 
accepts a query parameter as a SPARQL select statement and executes 
the parameter on (some configured?) SPARQL endpoints from looking at 
the single sourcefile I could find [1]. Having said that, I have been 
holding off getting my bio2rdf server to actually process rdf but it 
doesn't look so hard now. (The bio2rdf server is actually more generic 
than just biology or even bio2rdf but it is still named that in 
response to its origins. And in contrast to SQUIN it focuses on 
CONSTRUCT queries rather than SELECT)


On the subject of mashups I have been thinking in the last few days of 
combining the bio2rdf server with the pipes.deri.org 
http://pipes.deri.org interface for mashups, as some fairly 
sophisticated mashups can be done on pipes.deri.org 
http://pipes.deri.org, but a lot of the generic queries seem to be 
better handled at the client level where people can control with 
configurations what endpoints are used and have backups if a 
particular endpoint fails.


Cheers,

Peter

[1] http://tinyurl.com/6cvdl8

Peter,

Has anything happened re. cross-linking the data across bio2rdf.org and 
dbpedia.org?


Sane cross-linking is vital to Linked Data Web oriented Meshups.

Note, there is a distinct difference between a Mashup and a Meshup in my 
world view. Mashups are nice looking opaque Web pages that have code 
behind them while Meshups are transparent Web pages with Linked Data 
behind them (i.e. the data object URIs are accessible to machines). A 
Meshup style page is really the Linked Data Web's equivalent of a 
traditional DBMS View.



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: linked data mashups

2008-11-24 Thread Kingsley Idehen


Peter Ansell wrote:
2008/11/24 Kingsley Idehen [EMAIL PROTECTED] 
mailto:[EMAIL PROTECTED]



Peter Ansell wrote:

2008/11/23 Juan Sequeda [EMAIL PROTECTED]
mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
mailto:[EMAIL PROTECTED]


   Hi Giovanni and all


   On Sat, Nov 22, 2008 at 7:33 PM, Giovanni Tummarello
   [EMAIL PROTECTED]
mailto:[EMAIL PROTECTED]
   mailto:[EMAIL PROTECTED]
mailto:[EMAIL PROTECTED] wrote:


I guess that is THE question now: What can we do this
year
   that we
couldn't do last year?
( thanks to the massive amount of available LOD ).

   Two days ago the discussion touched this interesting
point. I
   do not
   know how to answer this question.
   Ideas?


   We need to start consuming linked data and making reall mashup
   applications powered by linked data. A couple of days I just
   mentioned the link for SQUIN: http://squin.sourceforge.net/

   The idea of SQUIN came out of ISWC08 with Olaf Hartig. The
   objective is to make LOD accesible easily to web2.0 app
   developers. We envision adding an S compoment to the LAMP
stack.
   This will allow people to easily query LOD from their own
server.

   We should have a demo ready in the next couple of weeks.

   We believe that this is something needed to actually start
using
   LOD and making it accesible to everybody.


How does SQIUN differ to a typical HTTP SPARQL endpoint? So
far it accepts a query parameter as a SPARQL select
statement and executes the parameter on (some configured?)
SPARQL endpoints from looking at the single sourcefile I could
find [1]. Having said that, I have been holding off getting my
bio2rdf server to actually process rdf but it doesn't look so
hard now. (The bio2rdf server is actually more generic than
just biology or even bio2rdf but it is still named that in
response to its origins. And in contrast to SQUIN it focuses
on CONSTRUCT queries rather than SELECT)

On the subject of mashups I have been thinking in the last few
days of combining the bio2rdf server with the pipes.deri.org
http://pipes.deri.org http://pipes.deri.org interface for
mashups, as some fairly sophisticated mashups can be done on
pipes.deri.org http://pipes.deri.org
http://pipes.deri.org, but a lot of the generic queries seem
to be better handled at the client level where people can
control with configurations what endpoints are used and have
backups if a particular endpoint fails.


Cheers,

Peter

[1] http://tinyurl.com/6cvdl8

Peter,

Has anything happened re. cross-linking the data across
bio2rdf.org http://bio2rdf.org and dbpedia.org http://dbpedia.org?


I have been waiting for information about what progress has been made 
with the community based infobox extraction framework. Then the 
relevant predicates in the protein/gene/chemical infoboxes can be used 
pretty easily for linkages.

So I am assuming this hasn't happened, based on your response?
 


Sane cross-linking is vital to Linked Data Web oriented Meshups.

Note, there is a distinct difference between a Mashup and a Meshup
in my world view. Mashups are nice looking opaque Web pages that
have code behind them while Meshups are transparent Web pages with
Linked Data behind them (i.e. the data object URIs are accessible
to machines). A Meshup style page is really the Linked Data Web's
equivalent of a traditional DBMS View.


I do understand the difference, but I tend to use the term mashup for 
any combining of the data sources independent of the presentation. Its 
hard enough defining a mashup when people ask for a definition without 
going for another similar term from my experience.
Labels are very secondary in how I tend to look at things. In due course 
the difference between Mashing and Meshing will be self evident, 
especially if we get all the major Linked Data hubs connected properly.


Kingsley


Cheers,

Peter




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: linked data hosted somewhere

2008-11-25 Thread Kingsley Idehen


रविंदर ठाकुर (ravinder thakur) wrote:

hi


i am trying to find some working service hosting the current 
linkedopendata and providing some programming interface. does any such 
service exists at the moment ?



thanks
ravinder thakur

Ravinder,

Could you be a little more specific? :-)

For instance, do you mean:

1. A service that provides SPARQL protocol access to all the warehoused 
data from the current LOD cloud?

2. ditto but using a service that is higher level than SPARQL?



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: linked cloud data hosted somewhere

2008-11-25 Thread Kingsley Idehen


Ravinder,
i am looking for a service which provide LOD with _any_ protocol. even 
a low level query system like sparql will be awesome!!!

Here is what we have in the making:

1. DBpedia SPARQL endpoint on EC2 (which simply means Virtuoso DB in the 
Clouds with DBpedia loaded, configured, and optimized) - ETA 24 hrs or 
less (we are just completing a number of reload runs so we have accurate 
sense of how long it takes to assemble DBpedia from a Virtuoso backup on 
EC2; current estimate is 1 1.5 hrs)


2. Rest of the LOD cloud warehouse in bits (i.e. one data space at a 
time like DBpedia with a sparql endpoint for each)


3. A Virtuoso Cluster edition with the entire LOD cloud warehouse pre 
loaded + sparql endpoint  (our Billion Triples Challenge demo [2] is the 
precursor of this deliverable).


We will announce the availability and provide installation guides for 
each of these linked data spaces in the clouds post AMI construction and 
commissioning.


Requirements:

1. Initialize an instance of the Virtuoso Universal Server paid AMI [2]


Links:

1. http://b3s.openlinksw.com/
2. 
https://aws-portal.amazon.com/gp/aws/user/subscription/index.html?offeringCode=6CB89F71


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: linked data hosted somewhere

2008-11-26 Thread Kingsley Idehen


Hugh Glaser wrote:

Thanks, Kingsley and Aldo.
I have to say you raise quite a lot of concerns, or at least matters of
interest.
I really don't think it is a big deal that I asked someone to consider
resources when accessing my web site, and I am a bit uncomfortable that I
then get messages effectively telling me that my software is poor and I
should be using (buying?) something else.
  

Hugh,

You're losing me a little, I don't think Aldo or I were making any 
comments about your software per se. or making suggestions about 
alternatives.


Anyway more comments inline below.

On 26/11/2008 02:12, Kingsley Idehen [EMAIL PROTECTED] wrote:

  

Hugh Glaser wrote:


I thought that might be the answer.
So what is the ontology of the error, so that my SW application can deal with
it appropriately?
If it ain¹t RDF it ain¹t sensible in the Semantic Web.
;-|
And the ³entitlement² to spend lots of money by accident; a bit worrying,
although I assume there are services that allow me to find out at least
estimates of the cost.

  

If you are querying via iSQL or the Virtuoso Conductor you wont be
moving lots of data between your desktop and EC2. If you do large
constructs over the sparql protocol or anything else that produces large
HTTP workloads between EC2 and your location, then you will incur the
costs (btw - Amazon are quite aggressive re. the costs, so you really
have to be serving many client i.e., offering a service for costs being
a major concern).


Er, yes, that was the question we were discussing.
Large constructs over the sparql prototcol.
With respect to costs, I never mentioned Amazon, so I am not sure why that
is the benchmark for comparison.
But I don't want to have a go at the Openlink software (I often recommend it
to people); I was just asking about limitations.
All software has limitations.
  

Anyway, Virtuoso let's you control lots of things, including shutting
down the sparql endpoint. In addition, you will soon be able to offer
OAuth access to sparql endpoint etc..


Yes, and I didn't really want to have the overhead of interacting with
Ravinder to explain why I had shut down his access to the SPARQL endpoint.
  

I suspect that your comment about a bill is a bit of a joke, in that normal
queries do not require money?
But it does raise an interesting LOD question.
Ravinder asked for LOD sets; if I have to pay for the query service, is it
LOD?

  

You pay for traffic that goes in and out of your data space.

(effective November 26, 2008)
Fixed Costs ($)


snip amazon costs/
  

Here is a purchase link that also exposes the items above.
https://aws-portal.amazon.com/gp/aws/user/subscription/index.html?ie=UTF8offe
ringCode=6CB89F71

Of course, you can always use the Open Source Edition as is and
reconstruct DBpedia from scratch, the cost-benefit analysis factors come
down to:

1. Construction and Commissioning time (1 - 1.5 hrs vs 16 - 22 hrs)
2. On / Off edition variant of live DBpedia instance that's fully tuned
and in sync with the master


Getting back to dealing with awkward queries.
Detecting what are effectively DoS attacks is not always the easiest thing to
do.
Has Bezzo really solved it for a SPARQL endpoint while providing a useful
service to users with a wide variety of requirements?

  

I believe so based on what we can do with Virtuoso on EC2.  One major
example is the backup feature where we can sync from a Virtuoso instance
into S3 buckets. Then perform a restore from those buckets (what we do
re. DBpedia). In our case we offer HTTP/WebDAV or the S3 protocol for
bucket access.


I don't think this contributes to helping to service complex SPARQL queries,
or have I missed somthing?
  


Hugh:  I certainly had my response above a little tangled :-(

To clarify, re.  Bezos and DOS bit. 


1.  EC2 instances can be instantiated and destroyed at will
2.  Virtuoso (and I assume other SPARQL engines) have DOS busting 
features such as SPARQL Query Cost Analysis and Rate Limits for HTTP 
requests.




In fact, people don¹t usually offer open SQL access to Open Databases for
exactly this reason.
I like to think the day will come when the Semantic Web is so widely used
that we will have the same problem with SPARQL endpoints.

  

The Linked Data Web is going to take us way beyond anything SQL could
even fantasize about (imho). And one such fantasy is going to be
accessible sparql endpoints without bringing the house down :-)


Now there I agree.
The power of LD/SW or whatever you call it will indeed take us a long way
further.
And I agree on the fantasy, which is actually what I was saying all along.
It is a fantasy to suggest that you can do all the wrong you want.
  

Exactly!

But I think it is sensible to take the question to a new thread...
  


No problem :-)


Kingsley

Best
Hugh
  

Kingsley




  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web

Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)

2008-11-26 Thread Kingsley Idehen
 reply.

My point was simply that, if you want to push things over the edge,
why not get your own box. We all take care of our infrastructure and
know its limitations.

So, I formally apologize.
I am by no means endorsing one piece of software over another ( save
for mine, but it does't exist yet ;).
My preferences for virtuoso come from experiential bias.

I hope this clears things up.
I apologize for the traffic.

However, I do make a formal request for some sense of humor.
This list tends to get into this kind of discussions, and we will
start getting more and more visits from outsiders who are not used to
this sort of sharpness.

Best,
A






  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)

2008-11-26 Thread Kingsley Idehen
://quebec.bio2rdf.org/download/virtuoso/indexed/
[3] http://quebec.bio2rdf.org/download/n3/
[4] http://sourceforge.net/project/platformdownload.php?group_id=142631
[5] http://bio2rdf.mquter.qut.edu.au/

--~--~-~--~~~---~--~~
You received this message because you are subscribed to the Google 
Groups bio2rdf group.

To post to this group, send email to [EMAIL PROTECTED]
To unsubscribe from this group, send email to 
[EMAIL PROTECTED]
For more options, visit this group at 
http://groups.google.com/group/bio2rdf?hl=en

-~--~~~~--~~--~--~---




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)

2008-11-29 Thread Kingsley Idehen


Danny Ayers wrote:

%=profanity /%
or something - cool thread

while I disagree with many of Aldo's individual points, getting them
surfaced is really positive

in response to a line from the firestarter:

The only reason anyone can afford to offer a SPARQL endpoint is because it
doesn't get used too much?

while my love of SPARQL is enormous, I can't see the SPARQL endpoint
being a lasting scenario.
  
But you seem to assume that SPARQL endpoints are the definitive scenario 
today?


I see SPARQL endpoints as simply optional. The public visibility will be 
determined by needs and the wishes of a give Linked Data Space owner.


Take DBpedia as an example, it works fine as a Linked Data Space on its 
own, but it also offers a SPARQL endpoint for those who want to speak 
SPARQL.


As I've stated (already), the real issue comes down to deep DBMS level 
issues re. scalability and query processing prowess. Today, we don't 
have a ODBC or JDBC DSNs published on the Web primarily because the 
naming granularity isn't there (i.e. record level naming), single record 
access and manipulation has always been a problem for SQL, and there's 
never been a truly standardized data access mechanism that incorporates 
the networking layer.


Linked Data picks up from where the RDBMS realm simply runs out of steam 
re. Open Data Access, but this doesn't mean that the techniques in the 
RDBMS realm can form the basis of a new kind of DBMS frontier of the 
kind I believe the Linked Data Web ultimately delivers.

linking and the fresh approach to caching this will demand, need
another rev. before the web starts doing data efficiently
  

When you say: ..Web doing data efficiently, what do you mean?

Scalable and change sensitive data access on the Linked Data Web is 
fundamentally an issue for Linked Data Deployment platform player 
differentiation (imho). I think HTTP, RDF, and SPARQL (plus extensions) 
already provide ample building blocks for building solid platforms. We 
just have to remember that data access and DBMS technology predates the 
Web and lots of the techniques and knowledge from this realm are vital 
when it comes to  scalability and efficiency challenges.


In my eyes, and from my experience, the Web is only beginning to 
comprehend DBMS matters. And like most things relating to humans, some 
lessons are only learned the hard way :-)


Kingsley


the answer to the quoted line is the question - how can you not
afford? Classic stuff re. amazon opening up their silo a little bit -
guess what, profit!

pip,
Danny.


2008/11/28 Juan Sequeda [EMAIL PROTECTED]:
  


On Thu, Nov 27, 2008 at 2:33 PM, Peter Ansell [EMAIL PROTECTED]
wrote:


2008/11/27 Richard Cyganiak [EMAIL PROTECTED]
  

Hugh,

Here's what I think we will see in the area of RDF publishing in a few
years:

- those query capabilities are described in RDF and hence can be invoked
by tools such as SQUIN/SemWebClient to answer certain queries efficiently


I still don't understand what SQUIN etc have that goes above
Jena/Sesame/SemWeb etc which can do this URI resolution with very little
programming knowledge in custom applications.
  

True, jena/sesame does everything that SQUIN entails to do. However, SQUIN
is oriented to the web2.0 developers. How is a php/ror web developer going
to interact with the web of data and make some kind of semantic-linked data
mashup over a night? SQUIN will let them do this. No need of having jena,
learning jena, etc. Make it simple! If it is not simple, then developers are
not going to use it.









  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: DBpedia+Virtuoso EC2 AMI

2008-12-01 Thread Kingsley Idehen


Dan Brickley wrote:

Kingsley Idehen wrote:


All,

Quick FYI.

What is Virtuoso+DBpedia AMI for EC2?
An pre-installed and fully tuned edition of Virtuoso that includes a 
fully configured DBpedia instance on Amazon's EC2 Cloud platform.


Nice work :)


Benefits?
Generally, it provides a no hassles mechanism for instantiating 
personal, organization, or service specific instances of DBpedia 
within approximately 1.5 hours as opposed to a lengthy rebuild from 
RDF source data that takes between 8 - 22 hours depending on machine 
hardware configuration and host operating system resources.


 From a Web Entrepreneur perspective it offers all of the generic 
benefits of a Virtuoso AMI on EC2 plus the following:


1. Instant bootstrap of a dense Lookup Hub for Linked Data Web 
oriented solutions
2. No exposure to any of the complexities and nuances associated with 
deployment of de-referencable URIs (you have a local DBpedia replica)


Does this mean all the URIs are dbpedia.mysite.example.com rather than 
dbpedia.org/* ?


In http://kingsley.idehen.name/page/Linked_Data I see
'http://dbpedia.org/resource/Linked_Data' at the top of the page, but 
no  owl:sameAs to http://kingsley.idehen.name/page/Linked_Data

Template bug.

It should read:
About: http://kingsley.idehen.name/resource/Linked_Data
Then hook back to DBpedia via owl:sameAs and/or an some atrribution tiples.




The raw data links aren't working for me at the moment:
eg.

curl 
http://kingsley.idehen.name/sparql?default-graph-uri=http%3A%2F%2Fdbpedia.orgquery=DESCRIBE+%3Chttp://kingsley.idehen.name/resource/Linked_Data%3Eoutput=xml; 


?xml version=1.0 encoding=utf-8 ?
rdf:RDF xmlns:rdf=http://www.w3.org/1999/02/22-rdf-syntax-ns#; 
xmlns:rdfs=http://www.w3.org/2000/01/rdf-schema#;


Template bug.


etc., ... but I get the impression the RDF files use the main dbpedia 
URI space, while the HTML uses local references. If that's the case, 
how might this look in RDFa?
It should be de-referencing the 
http://kingsley.idehen.name/resource/Linked_Data URI locally since the 
DBpedia data is in a local graph with IRI http://dbpedia.org . The 
routes back to the main/public DBpedia corpus / space is to be via 
owl:sameAs or Attribution triples.


Re. RDFa we will can put this in the template once this first cut 
settles.  I want to get DBpedia replicated with ease for personal and 
service specific use as task 1 :-).


Kingsley


cheers,

Dan

3. Predictable performance and scalability due localization of query 
processing (you aren't sharing the public DBpedia server with the 
rest of the world)


Features:
1. DBpedia public instance functionality replica (re. RDF and (X)HTML 
resource description representations  SPARQL endpoint)
2. Local URI de-referencing (so no contention with public endpoint) 
and part of the Linked Data Deployment

3. Fully tuned Virtuoso instance for DBpedia data set hosting




1. http://kingsley.idehen.name/resource/Linked_Data
2. http://kingsley.idehen.name/resource/Entity-attribute-value_model
3. http://kingsley.idehen.name/resource/Hyperdata
4. http://kingsley.idehen.name/resource/Object_hyperlinking
5. http://kingsley.idehen.name/resource/Barack_Obama







--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Potential Home for LOD Data Sets

2008-12-04 Thread Kingsley Idehen


Hugh Glaser wrote:

Exciting stuff, Kingsley.
I'm not quite sure I have worked out how I might use it though.
The page says that hosting data is clearly free, but I can't see how to get at 
it without paying for it as an EC2 customer.
Is this right?
Cheers
  

Hugh,

No, shouldn't cost anything if the LOD data sets are hosted in this 
particular location :-)



Kingsley

Hugh


On 01/12/2008 15:30, Kingsley Idehen [EMAIL PROTECTED] wrote:



All,

Please see: http://aws.amazon.com/publicdatasets/ ; potentially the
final destination of all published RDF archives from the LOD cloud.

I've already made a request on behalf of LOD, but additional requests
from the community will accelerate the general comprehension and
awareness at Amazon.

Once the data sets are available from Amazon, database constructions
costs will be significantly alleviated.

We have DBpedia reconstruction down to 1.5 hrs (or less) based on
Virtuoso's in-built integration with Amazon S3 for backup and
restoration etc..  We could get the reconstruction of the entire LOD
cloud down to some interesting numbers once all the data is situated in
an Amazon data center.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Potential Home for LOD Data Sets

2008-12-04 Thread Kingsley Idehen


Hugh Glaser wrote:

Thanks for the swift response!
I'm still puzzled - sorry to be slow.
http://aws.amazon.com/publicdatasets/#2
Says:
Amazon EC2 customers can access this data by creating their own personal Amazon 
EBS volumes, using the public data set snapshots as a starting point. They can 
then access, modify and perform computation on these volumes directly using 
their Amazon EC2 instances and just pay for the compute and storage resources 
that they use.
  
Does this not mean it costs me money on my EC2 account? Or is there some other way of accessing the data? Or am I looking at the wrong bit?
  
Okay, I see what I overlooked: the cost of paying for an AMI that mounts 
these EBS volumes, even though Amazon is charging $0.00 for uploading 
these huge amounts of data where it would usually charge.


So to conclude, using the loaded data sets isn't free, but I think we 
have to be somewhat appreciative of a value here, right? Amazon is 
providing a service that is ultimately pegged to usage (utility model), 
and the usage comes down to value associated with that scarce resource 
called time.

Ie Can you give me a clue how to get at the data without using my credit card 
please? :-)
  
You can't you will need someone to build an EC2 service for you and eat 
the costs on your behalf. Of course such a service isn't impossible in a 
Numerati [1] economy, but we aren't quite there yet, need the Linked 
Data Web in place first :-)


Links:

1. http://tinyurl.com/64gsan

Kingsley

Best
Hugh

On 05/12/2008 02:28, Kingsley Idehen [EMAIL PROTECTED] wrote:



Hugh Glaser wrote:
  

Exciting stuff, Kingsley.
I'm not quite sure I have worked out how I might use it though.
The page says that hosting data is clearly free, but I can't see how to get at 
it without paying for it as an EC2 customer.
Is this right?
Cheers



Hugh,

No, shouldn't cost anything if the LOD data sets are hosted in this
particular location :-)


Kingsley
  

Hugh


On 01/12/2008 15:30, Kingsley Idehen [EMAIL PROTECTED] wrote:



All,

Please see: http://aws.amazon.com/publicdatasets/ ; potentially the
final destination of all published RDF archives from the LOD cloud.

I've already made a request on behalf of LOD, but additional requests
from the community will accelerate the general comprehension and
awareness at Amazon.

Once the data sets are available from Amazon, database constructions
costs will be significantly alleviated.

We have DBpedia reconstruction down to 1.5 hrs (or less) based on
Virtuoso's in-built integration with Amazon S3 for backup and
restoration etc..  We could get the reconstruction of the entire LOD
cloud down to some interesting numbers once all the data is situated in
an Amazon data center.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com












--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Exercise: LOD questions (R)†was ( Do we need another list(s)? )

2008-12-06 Thread Kingsley Idehen
do it
:).

I dont think its a coincidence that some of the smartest
people who
worked on semantic web now work for them.  (but of course there is
much more than a good idea for a successful business so they
might go
bust anyway obviously)


 A final quote people like me don't a) know about this and
b) don't
 understand how to use it once they do? I would say some
additional education
 is necessary to make this understood... i would also say
that in a broader
 sense the semantic web message has gotten lost under a mass
of acryonyms and
 theory

for a more articulate attept at an explanation of what happened i
agree a lot with this post
 http://inamidst.com/whits/2008/technobunkum by Sean Palmer

I dont think more education is needed Juan, one really
should teach
something if .. the answer is known else its called
brainstorming or
handwaving (according to weather you're in  good faith or not)

note that this is all but a bashing on the power of handling
loosely
structured data and RDF. I think on the other hand RDFa will
triumph
and so people will be probably making their own little
vucabolaries..
but starting from the web 2.0 approach and practical how do i
bring
visitors, how to do simple site to site integration use cases.

Giovanni






--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Potential Home for LOD Data Sets - is it Open?

2008-12-07 Thread Kingsley Idehen


Hugh Glaser wrote:

Thanks Kingsley.
Getting there, I think/hope.
So exactly what is the URI?
I run something like
select *
where
 {
   ?s ?p ZNF492
 }
and get back things like http://purl.org/commons/record/ncbi_gene/57615, but
these are not URIs in the Amazon cloud, and so if that is where I was
serving my Linked Data from, they are not right.
  

Hugh,

You are experimenting with a work in progress.

DBpedia on EC2 is the showcase for now.

http://kingsley.idehen.name/resource/Linked_Data will de-reference 
locally and produce triples that connect to 
http://dbpedia.org/resource/Linked_Data.


But note, http://kingsley.idehen.name is currently serving up 
neurocommons data while finish what I am doing with neurocommons.

Would it look something like
http://ec2-67-202-37-125.compute-1.amazonaws.com/record/ncbi_gene/57615 or
something else?
  

Yes, and like the DBpedia example link back to the public neurocommons URIs.

Or is it just that neurocommons is not offering resolvable URIs on the EC2
(if I understand the term), but they could switch on something (in
Virtuoso?) that would give me back resolvable URIs on Amazon?
  
The instance on EC2 will do what I stated above once we are done with 
the de-referencing rules construction and verification etc..



And I am also now wondering who pays when I curl the Amazon URI?
It can't be me, as I have no account.
  
The person or organization deploying the linked data pays the bill for 
the computing resources by amazon and the linked management and 
deployment services from Virtuoso.



It isn't the person who put the data there, as you said it was being hosted
for free.
  
Be careful here, the hosting issue was simply about an additional home 
for the RDF data set archives. The place from with Quad / Triple stores 
load their data en route to Linked Data publishing (via EC2 or somewhere 
else).  In the case of an EC2 AMI the cost of loading from an S3 Bucket 
into an EC2 AMI is minimal (if anything at all) since the data is in the 
same data center as the EC2 AMI.

I assume that it means that it must be the EC2 owner, who is firing up the
Virtuoso magic to deliver the RDF for the resolved URI?
  
Yes, and in the case of Virtuoso you simply have a platform that offers 
Linked Data Deployment with a local de-referencing twist while retaining 
links to original URIs (as per my DBpedia example).


Once I am done with neurocommons I'll temporary put DBpedia and 
Neurocommons on different ports at http://kingsley.idehen.name for demo 
purposes :-)



Kingsley

Best
Hugh

On 07/12/2008 03:34, Kingsley Idehen [EMAIL PROTECTED] wrote:

  

Hugh Glaser wrote:


Thanks Kingsley.
In case I am still misunderstanding, a quick question:

On 06/12/2008 23:53, Kingsley Idehen [EMAIL PROTECTED] wrote:
...

  

Linking Open Data Sets on the Web, is about publishing RDF archives with
the following characteristics:

1. De-referencable URIs



...

So if someone decides to follow this way and puts their Linked Data in the
Amazon cloud using this method, can I de-reference a URI to it using my
normal browser or curl it from my machine?

  

Hugh,

Absolutely!

For instance, a EC2 based instance of DBpedia will do the following:

1. Localize the de-referencing task (i.e. not pass this on to general
public instance of DBpedia)
2. Project triples that connect back to the http://dbpedia.org via
owl:sameAs (*this was basically what Dan was clarifying in our exchange
earlier this week*)

The fundamental goal is to use Federation to propagate Linked Data
(meme, value prop., and business models) :-)

btw - Neurocommons is a data set is now live at the following locations:

1. http://kingsley.idehen.name (*temporary as I simply used this to set
up the AMI and verify the entire DB construction process)
2. http://ec2-67-202-37-125.compute-1.amazonaws.com/ (*instance set up
by Hugh to double-verify what I did*)

Neurcommons takes about 14hrs+ to construct under the best of
circumstances. The process is now 1.15 hrs and you have your own
personal or service specific neurocommons database.

Next stop, Bio2Rdf :-)



Thanks.
Hugh



  

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com









  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: The next Internet giant: linking open data, providing open access to repositories

2008-12-07 Thread Kingsley Idehen


Shavkat Karimov wrote:
I think people wouldn't pay any fees to search the web. 

I don't believe I said or implied they would.

The FREEdom of
the Internet is why it is so powerful.

Nobody is disputing that.

 Also, while I agree with most
other Milton's statements, this new Internet giant might still be
Google, since they are already going to this direction in many ways.
Thanks for the insightful post!
  

Again, the issue of giant/behemoths is somewhat irrelevant (imho).

The issue is where and how the Web evolves.

I believe we are adding the property oriented Find of the Linked Data 
Web aspect to the full text pattern oriented Search of the Document 
Web aspect.  The Web is simply evolving and so will the value it 
delivers etc..


Federation is making a comeback courtesy of the Linked Data aspect of 
the Web :-)



Kingsley

Kindest regards,
Shavkat
http://www.seomanager.com


---
Monday, December 8, 2008, 12:14:05 AM, you wrote:


  

Sw-MetaPortal-ProjectParadigm wrote:

The next Internet giant company will be linking open data and 
providing open access to repositories, in the process seamlessly 
combining both paid for subscriptions, Creative Commons or similar 
license based or open source software schemes.


Revenues will be generated among other things from online advertising 
streams currently not utilized by Google or Yahoo!


In the big scheme of things this company will redefine the concept of 
internet search to provide access to deep(er) web levels of data and 
information for which users will be willing to pay an annual flat fee 
subscription.


Sound improbable? Non-profit organizations dedicated to providing 
global open access will soon start exploring just such business 
schemes to determine if it is feasible to fund and maintain the server 
farms, hard and software to do just that.


  

Milton,



  

Not improbable, that's what coming :-)



  
The only tweak of you statement would be this: there will be a 
federation of linked data spaces rather than a single behemoth :-)



  

Kingsley


Milton Ponson
GSM: +297 747 8280
Rainbow Warriors Core Foundation
PO Box 1154, Oranjestad
Aruba, Dutch Caribbean
www.rainbowwarriors.net
Project Paradigm: A structured approach to bringing the tools for 
sustainable development to all stakeholders worldwide

www.projectparadigm.info
NGO-Opensource: Creating ICT tools for NGOs worldwide for Project Paradigm
www.ngo-opensource.org
MetaPortal: providing online access to web sites and repositories of 
data and information for sustainable development

www.metaportal.info
SemanticWebSoftware, part of NGO-Opensource to enable SW technologies 
in the Metaportal project

www.semanticwebsoftware.org


  





  



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: The next Internet giant: linking open data, providing open access to repositories

2008-12-07 Thread Kingsley Idehen


Sw-MetaPortal-ProjectParadigm wrote:

Kinglsey,

I would very much like to know what it will take to start up just such 
a federation.


There are technical standards, legal issues of copyright and numerous 
others to explore.


Federation post Web 2.0 will occur in the same fashion centralization 
occurred during the Web 1.0 and 2.0 stages of Web evolution  i.e.,  
services / solutions deployment models.


If we take Web 2.0 as a example, the model is inherently central 
courtesy of a specific application of the SaaS (Software as a Service) 
deployment model. Google, Yahoo, Facebook, MySpace, and the litany of 
others  (from Web 1.0 to 2.0) all fit my characterization in this regard.


Federation will resurface by the emergence of personalized variations of 
the SaaS deployment model courtesy of cloud computing platform 
offerings  like
Amazon EC2 (and others). We will end up with Personal Portals that 
expose Metadata in RDF based Linked Data form (pun intended :-) ),  
which is something that our OpenLink Data Spaces platform (ODS) [1] has 
long facilitated (i.e., all your Web 2.0 silos are hooked into a virtual 
data space that exposes the data from those enclaves as RDF based Linked 
Data).


How do we go about designing a framework for the federation, i.e. a 
defining framework document and the requirements for each 
participating space?



See how we've done it via the ODS framework and programmers guide [2][3].


And to use analogies from the community of self-autonomous entities, 
will all spaces be equal like e.g. nations or will we have 
city-states, capitol districts, autonomous territories, provinces and 
the likes to identify the interrelatedness and dependency structure of 
the spaces.



All of the above, in a nutshell.


A federation presupposes common ideals, principles, standards, and 
intentions governed by a declaring, defining document and a 
constituting, instituting document.


Well we are a loose-ish federation of social beings in our human 
meta-space, we just haven't translated that completely to cyberspace.


I know of at least three spaces myself would like to join such a 
federation as soon as possible!


I think all the silo-ed members of social-networks and other Web 2.0 
solutions will make this happen in due course. The opportunity cost of 
futile resistance will force the silo owners to play ball (it always 
does eventually).


We have all the essential building blocks in place, who will step to 
the fore to create and form part of a council of founding fathers for 
just such an endeavor?



You [4] :-)


The exciting part of this process is that both consumers groups of 
information and producers/providers of information represented in such 
a space of linked data could form part of the federation.


By doing so you will create a much better feedback loop about what 
users would actually appreciate in functionality first, obviously 
building on the standards created by W3C et alii.


An exciting idea this federation!


Yes.

Links:

1. http://virtuoso.openlinksw.com/wiki/main/Main/Ods
2. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSProgrammersGuide
3. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtuosoOdsUbiquityTutorials

4. http://tinyurl.com/6lwfqj


Kingsley


Milton Ponson
GSM: +297 747 8280
Rainbow Warriors Core Foundation
PO Box 1154, Oranjestad
Aruba, Dutch Caribbean
www.rainbowwarriors.net
Project Paradigm: A structured approach to bringing the tools for 
sustainable development to all stakeholders worldwide

www.projectparadigm.info
NGO-Opensource: Creating ICT tools for NGOs worldwide for Project Paradigm
www.ngo-opensource.org
MetaPortal: providing online access to web sites and repositories of 
data and information for sustainable development

www.metaportal.info
SemanticWebSoftware, part of NGO-Opensource to enable SW technologies 
in the Metaportal project

www.semanticwebsoftware.org


--- On *Sun, 12/7/08, Kingsley Idehen /[EMAIL PROTECTED]/* wrote:

From: Kingsley Idehen [EMAIL PROTECTED]
Subject: Re: The next Internet giant: linking open data, providing
open access to repositories
To: [EMAIL PROTECTED]
Cc: semantic-web [EMAIL PROTECTED], public-lod
public-lod@w3.org
Date: Sunday, December 7, 2008, 7:14 PM

Sw-MetaPortal-ProjectParadigm wrote:
 The next Internet giant company will be linking open data and providing
open access to repositories, in the process seamlessly combining both paid 
for
subscriptions, Creative Commons or similar license based or open source 
software
schemes.
 
 Revenues will be generated among other things from online advertising

streams currently not utilized by Google or Yahoo!
 
 In the big scheme of things this company will redefine the concept of

internet search to provide access to deep(er) web levels of data and 
information
for which users will be willing to pay an annual flat fee subscription.
 
 Sound

Update: EC2 Editions of DBpedia 3.2 Neurocommons

2008-12-08 Thread Kingsley Idehen


All,

DBpedia 3.2 [1] and Neurocommons [2] Linked Data Spaces now exist on 
EC2. In both cases you can now commission personal or service specific 
renditions of one or both data spaces.  The time for commissioning each 
of these is 1.5 hrs (or less) compared to 8 -22 hrs from source for 
DBpedia and 1.15 hrs (or less) compared to 14 hrs or more for Neurocommons.


Examples:

1. http://kingsley.idehen.name/resource/Linked_Data -- DBpedia 3.2 as a 
personal rendition on EC2
2. http://kingsley.idehen.name:8890 - Neurocommons in personal rendition 
form (see example HTML based Linked Data Space exploration at: 
http://kingsley.idehen.name:8890/science/owl/sciencecommons/gene_record)


Installation Guide:

1. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIDBpediaInstall 
- DBpedia installation guide for EC2
2. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMINeuroCommonsInstall 
- Neurocommons installation guide for EC2


Next up, Bio2RDF :-)

Note: The entire LOD data set collection is being constructed in 
parallel for similar deployment on EC2.



Links:

1. http://dbpedia.org/About
2. http://neurocommons.org/page/Main_Page

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Commercial Product Announcements

2008-12-08 Thread Kingsley Idehen


Ian Davis wrote:

Hi all,

I'm not aware of any official policy on commercial posts to this list, 
but usual mailing list etiquette generally recommends that postings 
about commercial services and products should indicate their nature 
and give some indication of cost. I think this is even more important 
when the mailing list is focussed on a project with an open or 
free theme. Obviously, as I represent a vendor my preference is to 
allow commercial postings but I would also prefer that they are 
clearly labelled as such. Is this a policy that other members would 
like to see made explicit?


Ian
Ian there is no know policy for product announcements or job postings or 
anything else. I think we just collectively focus on on relevance. I 
think the relevance to Linked Data is simply what matters.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Commercial Product Announcements

2008-12-09 Thread Kingsley Idehen


Amy Stephen wrote:
This could degenerate into a religious argument in short order. There 
is simply no way to specifically regulate conversation without 
throttling knowledge. Since this list is primarily focused on the 
enabling the free flow of knowledge, it seems a counterproductive move.


If the list continues to focus on purpose and data and technologies 
and protocols and standards, and people try really, really hard to 
avoid self promotion of any variety as much as possible, and remember 
to be considerate of and trust one another, then things work.


In the final analysis, cooperation and trust are our only hope for 
just about anything.

Amen!

Kingsley



On Tue, Dec 9, 2008 at 6:25 AM, Kingsley Idehen 
[EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote:



David Wood wrote:

On Dec 8, 2008, at 8:57 PM, Kingsley Idehen wrote:

Ian Davis wrote:

Hi all,

I'm not aware of any official policy on commercial
posts to this list, but usual mailing list etiquette
generally recommends that postings about commercial
services and products should indicate their nature and
give some indication of cost. I think this is even
more important when the mailing list is focussed on a
project with an open or free theme. Obviously, as
I represent a vendor my preference is to allow
commercial postings but I would also prefer that they
are clearly labelled as such. Is this a policy that
other members would like to see made explicit?

Ian

Ian there is no know policy for product announcements or
job postings or anything else. I think we just
collectively focus on on relevance. I think the relevance
to Linked Data is simply what matters.



I agree, for what that is worth.  Clear identification of
commercial content is certainly appreciated, though.  Even
better would be identification capable of being filtered on.

Regards,
Dave



Dave,

So provide a template to clarify your suggestion. Ambiguity
doesn't solve anything bar leaving grounds for subjective innuendo.

Lets be as clear as possible since this is how objectivity is
established.

Also, for what it's worth, what is commercial content ?


-- 



Regards,

Kingsley Idehen   Weblog:
http://www.openlinksw.com/blog/~kidehen
http://www.openlinksw.com/blog/%7Ekidehen
President  CEO OpenLink Software Web: http://www.openlinksw.com









--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: Zemanta API: from text to LOD

2008-12-09 Thread Kingsley Idehen


Andraz Tori wrote:

Hi,

I am Andraz Tori, CTO at Zemanta. This is my first post to this list.
I've read the discussion about commercial announcements just a day ago,
so I hope this this announcement is relevant enough albeit it comes from
a for-profit company.
  

Welcome!

Relevant!

Please feel at home :-)

I'd like to announce that Zemanta today launched a semantic API that is
able to take plain text and disambiguate most important entities found
to their meanings in Linking Open Data (currently dbPedia, MusicBrainz,
Semantic CrunchBase and Freebase).
  

I can't find the API itself or API demos.

I've seen your illustrations etc..

As far as we know this is first such undertaking of this scale.
  
Not necessarily, but being first doesn't matter, it's the value 
delivered that is ultimately the key factor (imho) :-)

API is free to use up to 10k calls per day (default limit is 1k, email
needs to be sent to raise it). Response can be in XML, JSON or RDF/XML.



Developer info can be found at http://developer.zemanta.com and less
technical explanations at http://www.zemanta.com/api

I am very interested if this will make any new types of Mashups happen
  
I would say potentially interesting meshups with the output of other 
Linked Data aware services, for instance. Everything meshes with 
everything in this realm once the essence is understood.

Also I am interested in feedback on RDF/XML format/ontology.
  
I see ourselves a bit like a stargate portal from unstructured text into

the LOD, enabling LOD to be used in broader number of situations.

Comments welcome :)

  

Please send a link to a page that provides sample Linked Data example.

Also clarify if you offer a REST style service as I only see your 
language specific APIs in your document web collective re. APIs etc..


I am assuming you offer something that works like this:

1. You receive a chunk of blurb as input
2. You analyze (where extraction and disambiguation occurs)
3. Return an RDF Linked Data graph in one or more formats as part of you 
output




--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO 
OpenLink Software Web: http://www.openlinksw.com








Re: [Ann] OSM Linked Geo Data extraction, browser editor

2008-12-10 Thread Kingsley Idehen


On 12/10/08 1:25 PM, Sören Auer wrote:


Hi all,

We were working in the last weeks on bringing geo data derived from 
the marvelous OpenStreetMap project [1] to the data web. This work in 
progress is still far from being finished, however, we would like to 
share some first preliminary results:


* A *vast amount of point-of-interest descriptions* was extracted from 
OSM and published as Linked Data at http://linkedgeodata.org


* The *Linked Geo Data browser and editor* (available at 
http://linkedgeodata.org/browser) is a facet-based browser for geo 
content, which uses an OLAP inspired hypercube for quickly retrieving 
aggregated information about any user selected area on earth.


Further information can be also found in our AKSW project description:

http://aksw.org/Projects/LinkedGeoData

Thanks go to Sebastian Dietzold, Jens Lehmann, Sebastian Hellmann, 
David Aumueller and other members of the AKSW team for their 
contributions.


Merry Christmas to everybody from Leipzig

Sören


[1] http://openstreetmap.org


Soren at Co.,

Very cool!

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: objects to facts to links to LOD to ???

2008-12-16 Thread Kingsley Idehen


On 12/16/08 10:44 AM, Knud Hinnerk Möller wrote:


Hi,

I'm not sure about the discovery part - that seems to be more a 
meta-layer with data describing linked datasets (VoiD [1]). However, I 
agree that inference would have to be in layer 4 and up. Things like 
non-explicit object consolidation, vocabulary and ontology mapping, 
etc. will be necessary to integrate the data even beyond the point of 
explicit links.


@ravinder: I like the three layers you propose - is there a reference 
for that, or did you come up with it just now? I'd love to reference 
it in my PhD thesis!


Cheers,
Knud

Knud et al,

I think Ravinder has started the process of fixing the current Semantic 
Web layer cake :-) Which is a very good thing (imho, but not seeking a 
Layer Cake discussion explosion).


The tricky part is the interchangeable nature of Discovery and Trust 
in any such scheme layer-wise. For instance, do Discovery and Trust 
occupy Layers 4, 5 or either ? We ultimately want to reason against 
trusted data sources, but the serendipity quotient of discovery is a key 
factor re. the dynamic nature of trusted sources.


Since I am clearly thinking and writing (aloud) at the same time, I 
would suggest:


Layer 4 - Discovery (with high Serendipity Quotient)
Layer 5 - Trust (albeit inherently volatile)

Kingsley



[1] http://semanticweb.org/wiki/VoiD
On 15.12.2008, at 20:54, Juan Sequeda wrote:


Hi Ravinder,

Interesting points. I would say that Layer 4 would be discovery. 
Furthermore, this would be an inference layer that would allow 
discovery. After having linked data, applications will be able to 
discover new data. So maybe layer 4 and up are part of the 
applications that enable discovery. Anyways, that is what I see what 
the semantic web is about: discovery and serendipity.


Juan Sequeda, Ph.D Student
Dept. of Computer Sciences
The University of Texas at Austin
www.juansequeda.com
www.semanticwebaustin.org


On Sat, Dec 13, 2008 at 11:16 AM, रविंदर ठाकुर (ravinder thakur) 
ravindertha...@gmail.com wrote:

Hello Friends,


We have now reasonably good repository of linked data but i am not 
able to find any good application being made out of it. From the 
dicsussion on these mailing lists and other places, i got the feeling 
that everyone in semantic world thinks that semantic web is something 
big but unfortunately nobody is able to think of an application for 
_general public_ that can be based on semantic web data we have 
currently.



This led me to wonder that are there any more layers of information 
organization missing in the semantic web stack that are needed to 
generate good usable semantic web data which can be more useful for 
generating useful semantic apps . Few of the concepts i was thinking 
for upper layers to implement were like categorization, set 
formation(eg MIT passout in 2002), set intersection, etc. With these 
higher level layers I was hoping to build a system to find higher 
level relations say between places with low cancer rates and the food 
people eat there or any perticular gene found in such people.



The current semantic stack looks like this:
.
layer5 (???)
layer4 ( ???)
layer3 linked facts (isa(mars,planet) AND partof(mars,solarsystem) 
AND mass(mars,6566.1255kg))

layer2 facts (isa(mars,planet), isa(cow,animal))
layer1 objects (eg cow, mars, man)



Are there any other thoughts on the need of layers above the layer 3 
(linked data ) or these layers will be defined by the respective apps 
developers ? Even if there isn't any need, i would atleast like to 
have a discussion on the kind upper level layers we might need :)



thanks
ravinder thakur





-
Knud Möller, MA
+353 - 91 - 495086
Smile Group: http://smile.deri.ie
Digital Enterprise Research Institute
National University of Ireland, Galway
Institiúid Taighde na Fiontraíochta Digití
Ollscoil na hÉireann, Gaillimh






--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Bio2Rdf EC2 AMI rendition is now live

2008-12-17 Thread Kingsley Idehen


All,

Following along the lines of DBpedia [1] and NeuroCommons [2], we now 
have a rendition of Bio2Rdf.org [3] available on the Amazon EC2 Cloud.

Total commissioning time:
1. 20 minutes for Virtuoso backup transfer from S3 bucket to EC2 AMI
2. 1.15 hrs for Virtuoso backup restoration.

As per usual, there is a *temporary* instance  deployed via one of my 
personal data spaces on EC2 which includes a SPARQL endpoint and HTML 
based linked data browser pages.


Some sample HTML browser page links:

1.  http://kingsley.idehen.name/geneid:4421752
2.  http://kingsley.idehen.name/uniprot:A0A379
3. http://kingsley.idehen.name/citations:17309688
4. http://kingsley.idehen.name/isoforms:A0A379-1

Links:

1. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIDBpediaInstall
2. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMINeuroCommonsInstall
3. 
http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIBio2rdfInstall



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







LOD Data Sets RDF Dump URLs

2008-12-18 Thread Kingsley Idehen


All,

We are now ready to load the entire LOD cloud into a Virtuoso 6.0 
(Cluster Edition) instance that will be available to the public (in 
identical manner to what we currently offer re. DBpedia).


I would like everyone that has published an RDF data set archive, that 
seeks to have their data exposed by this instance, perform the following 
steps:


1. Go to: http://esw.w3.org/topic/DataSetRDFDumps
2. Add your data set to the table (if it isn't already listed) or 
correct erroneous entries

3. Add a URL entry to the Archive URL column
4. Add a Publisher URI to the Publisher / Maintainer column (* this 
URI will be used to construction Attribution Triples *)


If you don't have a URI, you can get one very quickly by registering at: 
http://community.linkeddata.org/ods (*just register and you end up with 
a URI for yourself *).



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Hyperdata Hypercard

2008-12-22 Thread Kingsley Idehen



All,

Please watch.
http://www.archive.org/details/CC501_hypercard .  Especially in relation 
to: http://www.youtube.com/watch?v=hAAwKxeoCrk


History is always a great tutor when journeying into the future.

Concepts are quite static relative to contextual application and 
comprehension in an innovation continuum.


The Web is an innovation continuum that provides new and improved 
context for applying and comprehending old concepts.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







CORRECTION: Hyperdata Hypercard

2008-12-22 Thread Kingsley Idehen


On 12/22/08 5:40 PM, Kingsley Idehen wrote:


All,

Please watch.
http://www.archive.org/details/CC501_hypercard .  Especially in 
relation to: http://www.youtube.com/watch?v=hAAwKxeoCrk


History is always a great tutor when journeying into the future.

Concepts are quite static relative to contextual application and 
comprehension in an innovation continuum.


The Web is an innovation continuum that provides new and improved 
context for applying and comprehending old concepts.


The Web is part of an innovation continuum that provides new and 
improved context for applying and comprehending old concepts.


was what I meant to say :-)

--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: Granular dereferencing ( prop by prop ) using REST + LinkedData; Ideas?

2008-12-29 Thread Kingsley Idehen


On 12/29/08 10:51 AM, Aldo Bucchi wrote:

Hi All,

I am in the process of LODing a dataset in which certain properties
are generated on the fly ( props derived from aggregate calculations
over the dataset, remote calls, etc ). I would like to let the clients
choose which of these expensive properties they need on demand and on
a granular level.

For example, lets say I am interested in knowing more about resource
http://ex.com/a.
Per LD conventions, dereferencing http://ex.com/a ( via 303 ) returns

http://ex.com/a  a ex:Thing ;
   rdfs:label a sample dynamic resource;
   ex:dynamic1 45567 .

The problem is that the value for ex:dynamic1 is very expensive to
compute. Therefore, I would like to partition the document in such a
way that the client can ask  for the property on a lazy, deferred
manner ( a second call in the future ).
The same is true for dynamic2, dynamic3, dynamic4, etc. All should be
retrievable independently and on demand.

* I am aware that this can be achieved by extending SPARQL in some
toolkits. But I need LOD.
* I am also aware that most solutions require us to break URI
obscurity by stuffing the subject and predicate in the uri for a doc.
* Finally, seeAlso is too broad as it doesn't convey the information I need.

Anyone came up with a clean pattern for this?
Ideas?

Something as simple as:
GET http://x.com/sp?s={subject}p={predicate} ---  required RDF

works for me... but... .
If possible, I would like to break conventions in a conventional manner ;)

Best,
A

   

Aldo,

We will have something to demonstrate re. this matter once we are done 
with the Xmas / New Year break in relation to the entire LOD data set 
being loaded into Virtuoso. In a nutshell, we are going to expose Quad 
Store stats using VoiD. The Client shouldn't have to calculate anything 
anymore, the stats should be part of the data exposed by the RDF store.



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: Granular dereferencing ( prop by prop ) using REST + LinkedData; Ideas?

2009-01-04 Thread Kingsley Idehen


On 1/4/09 9:28 AM, Yves Raimond wrote:


Hi Richard!

On 3 Jan 2009, at 12:23, Richard Cyganiak rich...@cyganiak.de wrote:



On 2 Jan 2009, at 23:20, Yves Raimond wrote:
snip

I proposed this solution:
http://simile.mit.edu/mail/ReadMsg?listName=Linking%20Open%20DatamsgId=20926 



And some refinements here:
http://simile.mit.edu/mail/ReadMsg?listName=Linking%20Open%20DatamsgId=20962 


snip

This discussion was indeed interesting :)
I now tend to think that linking to a separate document is a cleaner
way to go, but I am still concerned about auto-discovery. When you see
something like:

:New_York :morePersonsBornHere http://example.org/persons_nyc.rdf .

in the representation of :New_York, you still need a way to describe
the fact that :morePersonsBornHere links to a document holding lots of
:birthPlace properties. Just saying that :morePersonsBornHere
rdfs:subPropertyOf rdfs:seeAlso won't do the job properly - how can
you tell before retrieving the whole document?


Yves, the proposal above addresses this. There would be a triple:

   :birthPlace link:subjectListProperty :morePersonsBornHere .

This triple can be either directly in the :New_York description, or 
in the vocabulary (where you'd find it by dereferencing 
:morePersonsBornHere).


The triple tells clients that they should follow the 
:morePersonsBornHere link if they are interested in :birthPlace 
triples. So, autodiscovery is solved.




Yes, it does work, but only for simple property lists. What about 
find here persons born in NYC between 1945 and 1975 ?





But perhaps the approach I proposed when we discussed the void:example
property could work, in exactly the same way as in [1].

In the representation of :New_York, we could write something like 
(in N3):


http://example.org/persons_nyc.rdf void:example { :al_pacino
:birthPlace :New_York }.


N3 formulae cannot be expressed in RDFa or RDF/XML. How would you 
serialize this in practice?




As in the post I refered to: you can point to 
http://example.org/dataset-example.rdf where you put these example 
triples.




Then, a SPARQL query like the following could find the documents that
hold information about persons born in New York:

SELECT ?doc
WHERE {
?doc void:example ?g .
GRAPH ?g {
?person :birthPlace :New_York .
}
}

One of the good thing with this approach is that the patterns of
information held in the target document can be arbitrarily complex -


As far as I can remember, all the examples that people have given 
could be addressed with a simple property-based approach. Has anyone 
mentioned a use case that goes beyond looking for a single property? 
If not, then what does the additional complexity of this proposal buy 
us in practice?


The example mentioned in my post uses more than one property, or the 
exampl above.





“This is good because it can deal with arbitrary complexity” is a 
fallacy on the Web. Usually it translates as: “This is bad because it 
makes the common case harder.”


(I note that the situation here is different from what you described 
in [1]. There it was about annotations on a dataset level. Here it is 
about annotating links that occur within many or all individual 
documents of a dataset.)




A RDF document is a dataset, and can be described as such :-)


Yves,

Just to clarify language (I do agree with your point). An RDF document 
is a resource (of type/kind information resource) and as a result -- 
like all Things -- it can be formally described via its attributes and 
relationship properties :-)


Happy New Year!

Kingsley

Cheers!
y


Best,
Richard





and the only thing you have to do is to provide an example RDF graph,
holding something representative of what you put in that document.

Cheers!
y

[1] 
http://blog.dbtune.org/post/2008/06/12/Describing-the-content-of-RDF-datasets 







--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: Contd: UMBEL Effect re. Faceted Browsing

2009-01-09 Thread Kingsley Idehen


On 1/9/09 4:32 PM, Jason Kolb wrote:
That's cool stuff Kingsley.  Seeing the two next to each other (with 
and w/out Umbel) makes it easy to see the value that Umbel adds.

Jason,

Yes, which is why I took an oath (with myself) last year to only demo or 
write about UMBEL when I have a simple high impact demo :-)


More to come.

Happy New Year!

Kingsley


On Fri, Jan 9, 2009 at 4:25 PM, Kingsley Idehen 
kide...@openlinksw.com mailto:kide...@openlinksw.com wrote:



All,

One thing that should emerge from this server side faceted browser
demo is the impact of inferencing based on UMBEL ontology.

Remember, UMBEL provides binding between the OpenCyc Upper
Ontology and domain specific ontologies such as FOAF. Compare
items 12 for an example of the effect of this Ontology Meshup:

1.
http://linkeddata.uriburner.com/about/html/http://xmlns.com/foaf/0.1/Person
 (no UMBEL inferencing enabled)

2.

http://b3s.openlinksw.com/about/?url=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2FPersonsid=177

http://b3s.openlinksw.com/about/?url=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2FPersonsid=177
 (UMBEL inferencing enabled)

Note the expanse of the FOAF vocabulary's graph courtesy of UMBEL
linkage to other ontologies via  Equivalence, Subclass, and other
links (what Fred refers to as  domain explosion in some of his
blog posts).


-- 



Regards,

Kingsley Idehen   Weblog:
http://www.openlinksw.com/blog/~kidehen
http://www.openlinksw.com/blog/%7Ekidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com









--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: [ANN] t4gm.info: a Linked Data rendering of the Thesaurus of Graphic Materials

2009-01-15 Thread Kingsley Idehen


On 1/14/09 6:13 AM, Bradley Allen wrote:
t4gm.info http://www.t4gm.info is a simple, lightweight and DRY 
http://en.wikipedia.org/wiki/DRY Linked Open Data rendering in 
XHTML+RDFa of the Library of Congress http://www.loc.gov/index.html' 
Thesaurus for Graphic Materials 
http://lcweb2.loc.gov/pp/tgmhtml/tgmabt.html. Details about its use 
and implementation can be found on its help page 
http://www.t4gm.info/help. Comments to the undersigned are welcomed. 
- regards, BPA


Bradley P. Allen
http://bradleypallen.org
+1 310 951 4300

Bradley,

Very cool!!


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: RKBExplorer Datasets in LOD Cloud Diagram (was Re: FW: [ESW Wiki] Update of SweoIG/TaskForces/CommunityProjects/LinkingOpenData by TedThibodeauJr)

2009-01-19 Thread Kingsley Idehen
 and delete it. Any use of this e-mail by an 
unauthorised recipient is prohibited.

Talis Information Ltd is a member of the Talis Group of companies and is 
registered in England No 3638278 with its registered office at Knights Court, 
Solihull Parkway, Birmingham Business Park, B37 7YB.

__
This email has been scanned by the MessageLabs Email Security System.
For more information please visit http://www.messagelabs.com/email
__

 




--
Dr Tom Heath
Researcher
Platform Division
Talis Information Ltd
T: 0870 400 5000
W: http://www.talis.com/



   



--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







DBpedia Update and Pending Changeover

2009-01-20 Thread Kingsley Idehen


All,

We are about to transition DBpedia from a Virtuoso 5 instance over to a 
Virtuoso 6.x (Cluster Edition) instance  sometime this week after a few 
days of live testing against a temporary secondary instance [1].  The 
new release also includes our much vaunted server side faceted browsing 
[2] that leverages SPARQL aggregates and configurable interaction time 
for query payloads (*note the retry feature which increments the 
interactive time window to demonstrate full effects*).


In the process of conducting this effort, we discovered that the Yago 
Class Hierarchy hadn't been loaded (courtesy of some TBox subsumption 
work  performed by the W3C's HCLS project participants such as Rob Frost).


In addition, Virtuoso 6.x introduces TBox subsumption via new 
Virtuoso-SPARQL transitivity options as exemplified below using the V6.0 
instance at:

http://dbpedia2.openlinksw.com:8895/sparql (or /isparql).


-- Yago Subsumption based AlphaReceptors query
select ?y
from http://dbpedia.org/resource/classes/yago#
where {
{ select *
  where { ?x rdfs:subClassOf ?y . }
}
option (transitive,
t_distinct,
t_in (?x),
t_out (?y)
)  .
filter (?x = 
http://dbpedia.org/class/yago/AlphaReceptor105609111)

 }

-- Yago Subsumption based Receptors query
select ?x
from http://dbpedia.org/resource/classes/yago#
where {
{ select *
  where { ?x rdfs:subClassOf ?y . }
}
option (transitive,
t_distinct,
t_in (?x),
t_out (?y)
) .
filter (?y = http://dbpedia.org/class/yago/Receptor105608868)
 }


Links:

1. http://dbpedia.org/resource/DBpedia - test URI
2. http://dbpedia2.openlinksw.com:8895/fct/facet.vsp - server hosted 
faceted browser.


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







Re: studies about LD visibility

2009-01-28 Thread Kingsley Idehen


On 1/27/09 1:24 PM, Samuel Rose wrote:


Things that cannot be done without LD:

I don't think there is technically anything that explicitly cannot be 
done without Linking Open Data.

Samuel,

I profoundly believe that there are things that cannot be achieved 
without Linked Data.


Now before I progress, let be clear about Linked Data. We have:

1. Linked Data the meme from TimBL re. adding datum level linkage 
between resources on the Web (hyperdata) which is (imho) short for: 
Linked Data Web or Web of Linked Data


2. Linked Data in a much broader sense covering the application / 
incorporation of HTTP into the time-tested Data Access By Reference 
pattern (pointers) commonly used at the OS and DBMS levels to access and 
manipulate data.


Irrespective of 12 specifics, there is one common theme: integration of 
disparate data sources without application, operating system, or network 
level impediments.  Basically, untethered data access by reference 
courtesy of an HTTP based pointer system.


Note: The ubiquity of HTTP, combined with its inherent ability to 
negotiate  referent (what you point to with a URI) description  
representation, is the source of its potency.


For eons (computer industry time) in the distributed computing realm, 
the ability to reference and de-reference an object (datum) via a 
pointer that's functional across application, operating system, and 
network boundaries has been the holy grail ; many technologies have 
tried an failed re. this pursuit, and the stories are engraved in the 
technology tombstones of efforts such as: corba, cairo, copeland, 
opendoc, ole, and others.


I hope I've clarified what Linked Data's USP really is.

btw - Linking Open Data (LOD) is about a community that adheres to a set 
of best practices for publishing data on the Web in a manner that 
facilitate datum level linkage (hyperdata links) in addition to existing 
document level linkage (hypertext links).


--


Regards,

Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
President  CEO
OpenLink Software Web: http://www.openlinksw.com







  1   2   3   4   5   6   7   8   9   10   >