Re: IRC channel?
Mark Diggory wrote: Theres an IRC server at w3.org. Is there any interest in starting a chat room for discussing LoD? -Mark ~ Mark R. Diggory - DSpace Developer and Systems Manager MIT Libraries, Systems and Technology Services Massachusetts Institute of Technology +1 . -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: imdb as linked open data?
Hugh Glaser wrote: On 03/04/2008 12:41, Kingsley Idehen [EMAIL PROTECTED] wrote: Hugh Glaser wrote: ... Hugh Hugh, This is an example of many to come, where LOD needs to pitch the value of Linked Data to Information Publishers :-) I think they will ultimately publish and host their own RDF Linked Data once the intrinsic value is clear to them. And when there is also actual extrinsic value? :-) :-) But yes, and making it easy for them, possibly by actually doing it for them, is part of the bootstrap process. Of course :-) The thing I am trying to work out is exactly how to make the pitch that fits with their business model, and where their profit line might come from. This requires a serious understanding of the detailed business model for the company in question (which is not necessarily a skill the an academic SW researcher has!). Here is my pitch: Make yourself more discoverable on the Web via your data . We also have similar LOD installations for CORDIS (the EU funding agencies' DB), NSF (a US funding agency), EPSRC (a UK funding agency), and ACM, among others. We have now engineered them so that they can be moved to the Information Publisher if desired. Such organisations sometimes have it as part of their remit to publicise the results, so they should be easier to deal with, in theory. And they benefit in the same way, in line with goals of their Grants :-) If anyone has a ready conduit to the appropriate place in such organisations, we would be delighted to talk with them, showing them what might be done. Others: Same here :-) Kingsley -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: LOD Community Wiki Space
Hausenblas, Michael wrote: Kingsley, Great to hear this, thanks! My first Q (sorry, didn't dig deep into it yet): How can you restrict the access for a page to a community or individuals (at least for a certain time span), for example to work on a proposal or coordinate a paper or the like? We can do this via the Admin part of MediaWiki. I can make you an Administrator if you choose? Kingsley Cheers, Michael -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH http://www.joanneum.at/iis/ -- -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen Sent: Tuesday, April 08, 2008 7:01 PM To: public-lod@w3.org Cc: public-lod@w3.org Subject: LOD Community Wiki Space All, We know have a MediaWiki instance at: http://community.linkeddata.org/MediaWiki. We can (optionally of course) use this Wiki instance to collaborative work on LOD and Linked Data related collateral etc.. The setup is explained at: http://community.linkeddata.org/MediaWiki/index.php?VirtuosoWi ki:About -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: ANN: semanticweb.org
Dan Brickley wrote: (resending with right email address for www-rdf-calendar; sorry for noise) Ditto for LOD Community :-) Dan Brickley wrote: +cc: RDF calendar list Denny VrandeÄić wrote: Hello all, we are proud to announce the relaunch of semanticweb.org semanticweb.org is a Semantic MediaWiki installation that gathers data about tools, persons, concepts, events, organizations, and publications on, around, and about the Semantic Web. The site provides cool URIs (as certified by the cool URI note authors, thanks Richard and Leo!) for everything, a SPARQL endpoint, several exports, and much more. We will see how the data from semanticweb.org can be reused in your applications. If you are at the WWW currently, come to tomorrow's presentation of Kalpana, semanticweb.org and the newest developments on Semantic MediaWiki, during the Dev Track Sessions from 10:30-12:00. There will be opportunities to ask questions, raise issues, make wishes, discussion etc. Kalpana is a tool for the easy reuse of semantic data inside your website. There will be a more detailed mail next week, if you want more information before that, pass by tomorrow :) This is really sweet! Great to see that site come back to live. Wish I could've been in China ... if you folks have any trip-reports to share, do post them here. Now, on to the feature requests! * http://semanticweb.org/wiki/Events ... can we get iCal feeds? * OpenID would be lovely too. Any chance of that? Re OpenID, I generated a FOAF Group description from the OpenIDs of everyone who has edited wiki.foaf-project.org (hmm I should crontab it). It is a nice way of building a grassroots list of the OpenIDs of people active in the RDF/SWIG and FOAF community. We could also take such a list and poke around to find blog RSS feeds etc. to auto-generate 'planet' aggregators, lots of possible applications I think. Re Calendar stuff, see http://semanticweb.org/wiki/Talk:Events also for requests/suggestions. The Events page is great. And anything we can do to move conference CFPs into a more automation-friendly mode would be a great help to all our mailboxes I think. Again, nice work. I look forward to hearing more on what's planned... cheers, Dan -- http://danbri.org/ -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Linked Data talks at XTech in Dublin (6th-9th May)
Andrew Walkingshaw wrote: On Fri, May 2, 2008 at 10:13 AM, Chris Bizer [EMAIL PROTECTED] wrote: there are plenty of interesting Linked Data talks at XTech this year. For instance: and a complete track on Open Data http://2008.xtech.org/public/schedule/topic/23 If you'll excuse the blatant self-promotion, I'm speaking on this kind of thing too - http://2008.xtech.org/public/schedule/detail/527 9:45am Friday: Representing, indexing and mining scientific data using XML and RDF: Golem and CrystalEye (Andrew Walkingshaw, University of Cambridge I hope what I'm going to say will be interesting to Linked Data people. A large chunk of my presentation's about automated extraction of triples from a corpus of a few tens of thousands of sets of experimental data in crystallography, and some of the things you can do with those triples once you've got them. Hope to see some of you there! Best wishes, Andrew Walkingshaw (http://wwmm.ch.cam.ac.uk/blogs/walkingshaw/) Andrew, My Linked Data Space http://www.openlinksw.com/dataspace/[EMAIL PROTECTED]/weblog/[EMAIL PROTECTED]/1355 (as in my public Blog Space) has been updated accordingly :-) Others, you may notice the I am kinda preoccupied with Linked Data dog-fooding :-) In this case I've rehashed the mail from Chris in palatable Linked Data form via a blog post. I will be doing the same with the LODW presentations once I have a sense of how many are actually live via URIs and how many I need to export to the Linked Data Web (via my Briefcase http://community.linkeddata.org/dataspace/person/kidehen2 / WebDAV space at http://community.linkeddata.org/ods ) from the various mail attachments received in response to Chris' initial mailer. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: ESWC 2008 Linked Data Playground
David Huynh wrote: Hi Richard, If you look at this version http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008-rdf.html you'll see that the RDF/XML file is linked directly. So, there's pretty much zero cost in getting RDF into Exhibit (it took me about 5 minutes to put that exhibit up). Exhibit automatically routes RDF/XML files through Babel for conversion. In the first version, I did that conversion and saved the JSON output so that Babel won't get invoked every time someone views the exhibit. That's an optimization. Of course, Babel isn't perfect in doing the conversion. Here is an iPhone mockup version for the same exhibit: http://people.csail.mit.edu/dfhuynh/projects/eswc2008/iphone/iphone-exhibit.html I only tested it on Firefox and Safari. I think the back button functionality doesn't quite work that well, but you get the idea. David David, Even if you don't use RDFa to express what is about i.e. it's foaf:primarytopic, foaf:topic etc.. In the Exhibit pages head/ You can accompany: link href=http://data.semanticweb.org/dumps/conferences/eswc-2008-complete.rdf; type=application/rdf+xml rel=exhibit/data / with link rel=alternate href=http://data.semanticweb.org/dumps/conferences/eswc-2008-complete.rdf; type=application/rdf+xml / I think we need to adopt a multi pronged approach to exposing Linked Data (the raw data behind the Web Page): 1. Content Negotiation (where feasible) 2. link rel=.../ (for RDF sniffers/crawlers) 3. RDFa Re. point 2, I've just taken a random person Abhita Chugh http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fsemanticweb.org%2Fwiki%2FAbhita_Chugh from http://data.semanticweb.org which exposes the RDF based Description of Abhita Chugh http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fsemanticweb.org%2Fwiki%2FAbhita_Chugh via our RDF Browser without problems (we use all 3 of the methods above to seek Linked Data associated with a Web Document). In this case it also eliminates the need to translate anything (e.g. routing via Babel) since the original data source is actually RDF. Of course, I could take the exhibit page and slap this in myself, but I am hoping you could tweak Exhibit such that it does point 2 and maybe point 3 automatically. That would be a major boost re. Exhibit's Linked Data credentials :-) Kingsley Richard Cyganiak wrote: David, On 28 May 2008, at 22:01, David Huynh wrote: Nice data! I find it useful to look at it this way: http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008.html Very cool! Can you tell us a little bit about the process of getting from our RDF to this Exhibit? How much manual/expert work did it involve? (Exhibit is awesome for presenting structured data, and I wonder how hard it is in general to get from RDF to Exhibit.) Cheers, Richard Good to see evaluation and interface big enough in the abstract word cloud. :-) David Chris Bizer wrote: Hi all, Paul, Richard, Knud, Tom, Sean, Denny and I have published some data describing papers and authors of the 5th European Semantic Web Conference (ESWC2008). The data is interlinked with DBpedia, Revyu and the Semantic Web Community Wiki. So if you add a review about a paper to Revyu or if you add something to the wiki, your data will mix nicely with the data that is already published about the conference. See http://data.semanticweb.org/conference/eswc/2008/html for a description of the dataset and its use cases. The data is currently also being crawled by several Semantic Web Search Engines and we hope to be able to add a section about how to use the data within these Search Engines before the weekend. If you have any ideas for further use cases or know about other Semantic Web client applications that could be used to navigate, visualize or search the data please let me know so that we can add them to the webpage. Have fun playing with the data! Cheers Chris -- Chris Bizer Freie Universität Berlin +49 30 838 54057 [EMAIL PROTECTED] www.bizer.de -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Contd: ESWC 2008 Linked Data Playground
David Huynh wrote: Exhibit can't do #2 because it only acts on the page at runtime, so the author of an exhibit must put that in herself. And that I just did for the ESWC 2008 exhibits. David, Here are some OpenLink RDF Browser views of your Exhibit generated ESW 2008 :: Papers http://people.csail.mit.edu/dfhuynh/projects/eswc2008/eswc2008-rdf.html Page, based on your link rel=alternate/ tweak: 1. New Browser Edition View http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fpeople.csail.mit.edu%2Fdfhuynh%2Fprojects%2Feswc2008%2Feswc2008-rdf.html; 2. Old Browser Edition View http://demo.openlinksw.com/rdfbrowser/?uri=http%3A%2F%2Fpeople.csail.mit.edu%2Fdfhuynh%2Fprojects%2Feswc2008%2Feswc2008-rdf.html -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: More ESWC 2008 Linked Data Playground
David Huynh wrote: Kingsley Idehen wrote: David, But should such links be on these pages, too? Absolutely, in line with the ideal best practice, you should have the following routes: 1. Request RDF via Content-type in your HTTP request 2. link rel=-alternate/ 3. GRDDL profile in head/ plus link rel=transformation.../ 4. eRDF or RDFa If 1-4 aren't available, Construct a URI that passes the pages through an RDFization Service (Babel, Virtuoso Sponger Service, Zitgist Services, Triplr, others). Kingsley http://www.eswc2008.org/main_program.html http://www.eswc2008.org/program.html I did a View Source and couldn't find any alternate RDF/XML link. In fact, I can't find any RDF/XML on that site. Maybe I wasn't looking hard enough. David, Then the official ESWC2008 site (eswc2008.org) is not following that ideal best practice since it's missing 2, 3, and 4, and as a user with just a standard-compliant browser there's little I can do to verify 1. Is this diagnose correct? If so, will you please tell the eswc2008 site admin to get with the program? :-) Sure! I am hoping that key principals associated with ESWC 2008 are following this thread :-) We all love to see more open linked data, of course, so it'd be nice if *SWC conference sites all lead as examples. In doing so, maybe we'll discover the real motivations (or severe lack thereof) for publishing SW content, and the challenges of publishing (e.g., tedious manual or complex programmatic process to get lat/lng coordinates). Amen to Dog-fooding! I hope we are getting closer to the day when the dialog sample below becomes the norm: Technology Vendor or Proponent: I am a vendor and/or proponent of Technology X that unveils the virtues of a given paradigm e.g Linked Data Technology Customer: Do you exploit the virtues of the technology yourself? If so, please show me how. The scenario above is very different from the general practice which always omits the vital Dog-fooding aspect :-( Kingsley David -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: More ESWC 2008 Linked Data Playground
David Huynh wrote: Thanks, Kingsley. Your responses were crystal clear. But http://www.eswc2008.org/main_program.html is still devoid of RDF. Yes :-( Sean: What's going on with the event's linked data space? Kingsley David Kingsley Idehen wrote: David, To this specific point you made: I believe there need to be a mechanism for rewarding RDF publishing, or scolding for forgetting. Do you have that mechanism in-place? My response: 1. Reward for RDF publishing comes down to the benefits or opportunity costs associated making a structured data contribution to the Web. If you contribute structure you open up the possibility for collective participation by others in the Web Community (users and developers). If you don't, then you incur the opportunity costs of having to do it all yourself. Expanding my response is best done by reading some of my most recent posts about the emergence of structure on the Web in general etc.. 2. The mechanism ultimately comes down to degrees to which relevant things are discovered in a given space e.g spaces espousing the virtues of Linked Data should honor our best practices and radiate the values of Linked Data, if they don't, then at the very least, as a community, we can flag omissions etc.. I pinged you about a little tweak to you exhibit Now, if my responses are not in line with your question, and I am not at ESWC 2008, but absolutely honor the value of discourse driven problem resolution, please extend this conversation via some very specific suggestions etc.. What would you like to see in place in relation to the questions you've posed? Assuming my responses aren't clear enough? -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Geohash for events
Bernard Vatant wrote: Hi masa Hi Bernard, Added support for event page uri: http://www.kanzaki.com/ns/geo/u07t4qf8j:2008-09-28_2008-10-03;INRIA_IST_2008?uri=http://www.inria.fr/actualites/colloques/2008/ist08/ Really cool. The URI format looks perfect to me now and exactly what I imagined you would do :-) . The uri part cannot be combined 'hash:datetime;name' part, because such uri itself adds another hierarchy to the original uri (i.e. / in event page uri). Hence it should be provided as query string. I wonder this looks too complicated for practical use ? It is, if users have to concatenate the URI themselves by going to geohash, searching the place, copying the geohash id in the service namespace, add the time interval in conformant date format, add the event URI. Speak about user experience ... for geeks like you and me, but ordinary people will never do it that way. But it you (or someone else) provide a smart bookmarklet (Faviki has given me a lot of ideas for that matter) to use in your browser from the page of the event (http://www.inria.fr/actualites/colloques/2008/ist08/), where the user can call geohash via a geocoder, enter the dates using a calendar applet, and grab the name from the page title ... et voilà ... The service would return a page as yours, with the RDF description and a permanent URI. And maybe call au passage the geonames service to add the neighbouring geonames features, yahoo or google to add sponsored links, whatever ... This would be *practical* ... You could even au passage dump the created event in a backend data store, etc. Say what? URI are to be felt, experienced, rather than seen. Yes, the extensions, plugins, plain old anchor text, and the like are the way to go, the Geohase URI scheme are currently implemented is really cool! I've just opened an example URI in a Browser Session http://demo.openlinksw.com/rdfbrowser/?uri=http%3A%2F%2Fwww.kanzaki.com%2Fns%2Fgeo%2Fu07t4qf8j%3A2008-09-28_2008-10-03%3BINRIA_IST_2008%3Furi%3Dhttp%3A%2F%2Fwww.inria.fr%2Factualites%2Fcolloques%2F2008%2Fist08%2F:-) Note, that the Map and TimeLine control provide appropriate views. Kingsley 2008/6/9, Bernard Vatant [EMAIL PROTECTED]: One thing I was wondering was how to encapsulate the URI of the event itself, something like (completely incorrect syntax, but you get the idea again) cheers, -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Redirection strategy
Richard Light wrote: In message [EMAIL PROTECTED], Sergio Fernández [EMAIL PROTECTED] writes Hi, This is a typical abstract URI: http://collections.wordsworth.org.uk/object/GRMDC.C104.9 Is anyone able to diagnose what might be wrong with my page-serving strategy? It worked well for me, see Vapour's report [1] over that URI. That's very encouraging. What was the problem that you saw? Try: http://tinyurl.com/3w7wrf The RDF browser seems to load an RDF graph based on the HTML page, rather than the RDF. It can only be the redirection which is failing: if you give it the resource URL you get the correct set of triples. Richard Richard, The output below indicates we have something to look into re. our Browser. curl -I -H Accept: application/rdf+xml http://collections.wordsworth.org.uk/object/1993.25 HTTP/1.1 303 See Other Date: Thu, 12 Jun 2008 02:10:54 GMT Server: Microsoft-IIS/6.0 X-Powered-By: ASP.NET Location: http://collections.wordsworth.org.uk/object/data/1993.25 Connection: close Accept-Ranges: bytes Vary: accept Content-Length: 0 Content-Type: application/rdf+xml;qs=1 Set-Cookie: ASPSESSIONIDSCSAQRAR=FAKPBAEADBPFBFAMOPLBPOAM; path=/ Cache-control: private -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: bbc-programmes.dyndns.org
Nicholas Humfrey wrote: hello, I am trying to get the work we did on: http://bbc-programmes.dyndns.org/ live on bbc.co.uk. Does anyone think that there anything that needs changed/fixed before it does go live? At the moment we just have RDF/XML views for Brands/Series/Episodes and Versions. But plan to is to also have RDF views for for the aggregation pages (tags, genres, formats, services, schedules...) some time in the future. It seems to be hard to find a consensus on use of URIs, but here is how things are the at moment. HTML Document: http://www.bbc.co.uk/programmes/b00b07kw RDF Document: http://www.bbc.co.uk/programmes/b00b07kw.rdf The thing: http://www.bbc.co.uk/programmes/b00b07kw#episode When asking for RDF here: http://www.bbc.co.uk/programmes/b00b07kw#episode you get 303 redirected here: http://www.bbc.co.uk/programmes/b00b07kw.rdf Description of Episode I assume is: http://www.bbc.co.uk/programmes/b00b07kw (an RDF based information resource i.e b00b07kw.rdf without the .rdf extension) The use of # cuts out the extra hops associated with a 30X :-) Example: http://kidehen.idehen.net/dataspace/person/kidehen#this The information resource hosting the description of This Entity (#this) is at: http://kidehen.idehen.net/dataspace/person/kidehen Hope this helps. Kingsley Is that sane, or just it infact be a 302? nick. http://www.bbc.co.uk/ This e-mail (and any attachments) is confidential and may contain personal views which are not the views of the BBC unless specifically stated. If you have received it in error, please delete it from your system. Do not use, copy or disclose the information in any way nor act in reliance on it and notify the sender immediately. Please note that the BBC monitors e-mails sent or received. Further communication will signify your consent to this. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: fans of Linked Open Data and the NFL, NHL, or Major League Baseball?
Bob DuCharme wrote: Kingsley Idehen wrote: Today we have ESPN and co. offering analysis via the traditional one-way TV medium. Tomorrow, I envisage a conversation space connected by analytic insights from Joe Public the analyst. Also, what's good for sports applies to Politics, Finance, Soap Operas, and other realms. What I'd really like to do as a next step is to identify a Linked Open Data advocate who happens to watch a lot of one of the sports for which more data is becoming available, and then encourage that person to get in touch with one of the stats-oriented communities within that sport to help make these stats available as a SPARQL endpoint. If it was soccer, Uche would be an obvious candidate, but judging by http://www.sportsstandards.org/oc the NHL, NFL, or MLB seem to be the best places to start. At the next Boston/Cambridge/semweb/LOD meetup, I suggest you keep you eye out for Red Sox, Bruins, or Pats-themed clothing... Bob Bob, I am a major Soccer, Football (NFL), Basketball (NBA), and Baseball (MLB) (play-offs only) fan. Uche and I share Soccer fan DNA amongst other things :-) I've looked at the site you suggested and it provides a number of great data sources for LOD and RDB2RDF projects. We will certainly have a crack at a Linked Data Space based on what's available. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Unobtrusive Manifestation of the Linked Data Web
All, In an attempt to kill many birds with a single stone here is a #swig logger session link http://chatlogs.planetrdf.com/swig/2008-06-22#T20-23-40 covering the thinking behind a new extension re. are realising for Firefox 2.x and 3.x (of course other browsers to follow). Gist of the matter: We are adding: View | Linked Data Sources, to browsers, so that Linked Data is revealed to Web Users as the 3rd of the following options: 1. View Rendered Page (what you get by default) 2. View Page Source (how you look at the markup behind the page) 3. View Linked Data Sources (how you look at the raw data behind the page; the data the page puts in context in rendered form) Of course, the same applies to the Browser Context Menus, meaning: Linked Data Sources occurs as an option (for completeness this will be changed to: View Linked Data Sources). Our extension uses an HTTP Proxy (Sponger) which is associated with RDFizers (think of these as RDF analogs of ODBC/JDBC Drivers, but for Web Data). Thus, you can increase or decrease the quality of the linked data graph via the Cartridges/Drivers that you plug into your data spaces. Also note, the recent enhancement to Semantic Radar now makes it possible to incorporate Linked Data Browsers/Viewers from a myriad of providers (OpenLink, Zitgist, Marbles, DISCO, Tabulator etc.), which is great, but you will note that the new plugin doesn't mandate any discovery of RDF. It simply takes a URL and then RDfizes it whenever you take the: View | Linked Data Sources menu or context menu route. The initial cut of the extension is at: http://myopenlink.net:8890/~kidehen/Public/rdfb.xpi Of course, there are a few updates on the way later this week(mostly aesthetic bar the non functioning Search feature). That said, you can get to the essence of what I am talking about via the initial release. Related: 1. My Linked Data Planet Keynote http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2.html (thanks to RDFa and the Bibliographic Ontology it's now possible to model Presentations (Slideshows) ) 2. Other Slidy Presentations from OpenLink http://virtuoso.openlinksw.com/Whitepapers/index.htm enhanced with RDFa annotations. Final note: since we already have a Cartridge/Driver for Slidy based data sources, you can also view the data behind any Slidy presentation. Of course, the context fidelity of the graph would be low (i.e. low linked data resolution) without RDFa or some other Linked Data exposure mechanism, but if you have a meta cartridge (e.g. the kind that Zitgist delivers via it's exploitation of UMBEL aided Named Entity Extraction and Disambiguation), you don't even require RDFa en route to context fidelity (HIgh Def. Linked Data), you just slot that cartridge into your data space. BTW - http://community.linkeddata.org/ods remains open to community members seeking both a data space and an easy place to get an Entity URI :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Contd: Unobtrusive Manifestation of the Linked Data Web
All, Quick note re. the Firefox Linked Data Sources extension ( rdfb.xpi) : 1. Download to your local filesystem 2. Use File | Open from Firefox to load (this isn't a properly signed XPI at the current time) Kingsley All, In an attempt to kill many birds with a single stone here is a #swig logger session link http://chatlogs.planetrdf.com/swig/2008-06-22#T20-23-40 covering the thinking behind a new extension re. are realising for Firefox 2.x and 3.x (of course other browsers to follow). Gist of the matter: We are adding: View | Linked Data Sources, to browsers, so that Linked Data is revealed to Web Users as the 3rd of the following options: 1. View Rendered Page (what you get by default) 2. View Page Source (how you look at the markup behind the page) 3. View Linked Data Sources (how you look at the raw data behind the page; the data the page puts in context in rendered form) Of course, the same applies to the Browser Context Menus, meaning: Linked Data Sources occurs as an option (for completeness this will be changed to: View Linked Data Sources). Our extension uses an HTTP Proxy (Sponger) which is associated with RDFizers (think of these as RDF analogs of ODBC/JDBC Drivers, but for Web Data). Thus, you can increase or decrease the quality of the linked data graph via the Cartridges/Drivers that you plug into your data spaces. Also note, the recent enhancement to Semantic Radar now makes it possible to incorporate Linked Data Browsers/Viewers from a myriad of providers (OpenLink, Zitgist, Marbles, DISCO, Tabulator etc.), which is great, but you will note that the new plugin doesn't mandate any discovery of RDF. It simply takes a URL and then RDfizes it whenever you take the: View | Linked Data Sources menu or context menu route. The initial cut of the extension is at: http://myopenlink.net:8890/~kidehen/Public/rdfb.xpi Of course, there are a few updates on the way later this week(mostly aesthetic bar the non functioning Search feature). That said, you can get to the essence of what I am talking about via the initial release. Related: 1. My Linked Data Planet Keynote http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2.html (thanks to RDFa and the Bibliographic Ontology it's now possible to model Presentations (Slideshows) ) 2. Other Slidy Presentations from OpenLink http://virtuoso.openlinksw.com/Whitepapers/index.htm enhanced with RDFa annotations. Final note: since we already have a Cartridge/Driver for Slidy based data sources, you can also view the data behind any Slidy presentation. Of course, the context fidelity of the graph would be low (i.e. low linked data resolution) without RDFa or some other Linked Data exposure mechanism, but if you have a meta cartridge (e.g. the kind that Zitgist delivers via it's exploitation of UMBEL aided Named Entity Extraction and Disambiguation), you don't even require RDFa en route to context fidelity (HIgh Def. Linked Data), you just slot that cartridge into your data space. BTW - http://community.linkeddata.org/ods remains open to community members seeking both a data space and an easy place to get an Entity URI :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Linked Data re Non-Profits and NGO's. Have data, need vocabulary.
Chris Bizer wrote: Hi Bob and Kingsley, a while ago there was a RDF version of Wikicompany online at http://dbpedia.openlinksw.com/wikicompany/resource/Wikicompany Maybe it would also be an idea to reuse terms from the vocabulary of this source. Kingsley: The URI abouve currently gives a 500 return code. Do you know what happend to the site? As an alternative you could also think about reusing the terms that are currently used within DBpedia to describe organizations. Chris, Wikicompany is certainly the place to start. The URI should be live again. Kingsley Cheers Chris -- Chris Bizer Freie Universität Berlin +49 30 838 54057 [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] www.bizer.de http://www.bizer.de - Original Message - *From:* Bob Wyman mailto:[EMAIL PROTECTED] *To:* public-lod@w3.org mailto:public-lod@w3.org *Sent:* Friday, June 27, 2008 1:19 AM *Subject:* Linked Data re Non-Profits and NGO's. Have data, need vocabulary. I would like to make available as Linked Data several databases describing several million non-profits, NGO's and foundations. The data includes things like name of organization, address, budget, source of funds, major programs, key personnel, relationships to other organizations, area of expertise, etc. What I don't have is an RDF vocabulary with which to describe these things. While I could define one myself, I would like to base my work on existing standards, or common practice, however, seemingly endless digging through the web indicates that there aren't any obvious standards for describing even basic things like address in RDF. Perhaps, I'm looking in the wrong places... Ideally, I would find some well formed vocabulary for a Business Description that I could use or adapt. I would appreciate it if anyone could give me pointers to either such a well worked vocabulary or at least to smaller vocabularies for things like address that I could use in composing a vocabulary with which to publish this data. Can you help? bob wyman -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)
Tom Heath wrote: -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen Sent: 12 July 2008 21:43 To: [EMAIL PROTECTED] Cc: public-lod@w3.org; [EMAIL PROTECTED] Subject: Re: Ordnance Survey data as Linked Data (RE: How do you deprecate URIs? Re: OWL-DL and linked data) Kingsley Idehen wrote: I also forgot to mention obvous use of RDFa in the HTML doc which broadens the range of rdf aware user agents tha commence RDF discovery from HTML Question: is it worth creating a duplicate RDF graph by using RDFa in HTML documents, when there is also RDF/XML available just one link rel=.../ away, and at a distinct URI? Doesn't this RDFa + RDF/XML pattern complicate the RDF-consumption picture in general if we assume agents will want to do something with data aggregated from a number of sources/locations, i.e. doesn't it increase the cost of removing duplicate statements by creating more in the first place? Does it not also complicate the picture of making provenance statements using named graphs, if the subject of the triple could be both an HTML document and an RDF graph? Dunno the answers to these questions, but interested to hear what people think. Tom. Tom, I believe we should spread the net as wide as possible re. RDF aware user agents since we cannot assume which methods they will be using to discover RDF. On our part (re. our RDF aware user agents), we use all the methods (Content Negotiation, link /, GRDDL, RDFa, and POSH) when sniffing, which also implies that we take on the burden of normalization. On the part of publishers, I encourage the use of at least one of the RDF exposure methods mentioned above, when known associations between HTML and RDF representations exist. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)
Hugh Glaser wrote: Thanks Tom. Er, yes. I was puzzled by the suggestion that I might duplicate the RDF in the page that did a simple html rendering of the underlying RDF I was trying to publish. I would have thought that this is actually a Bad Thing, rather than a Good Thing. And if we are talking about an RDF browser (as our pages are, albeit with a clean URI that doesn't have the browser URI in it), getting it to include the RDF as RDFa or whatever is even stranger; after all http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fos.rkbexplorer.com%2Fdescription%2Fosr70017765 doesn't include the substantive RDF as RDFa, (or have a link rel to http://os.rkbexplorer.com/data/osr70017765 for that matter) which would be the equivalent. Hugh, I knew this was coming. Please practice what I say and not the current state of our user agents :-) I prefer to make suggestions that go beyond what we've implemented. Yes, of course, I advocate dog-fooding but I also have to deal with the realities of development and product release cycles etc.. Our agents will be fixed in line with my suggestions, for sure. My key point is this: we are a community of knowledgeable folks who need to take on the the burden of unobtrusive injection of RDF into the Web. We cannot expect this to happen outside the community at this stage. Using HTML as the vehicle for RDF exposure is a big deal and offers immense value. RDFa is a major contribution to the whole RDF exposure puzzle, and this is one area where it's value is crystal clear (imho). Kingsley On 14/07/2008 09:55, Tom Heath [EMAIL PROTECTED] wrote: -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen Sent: 12 July 2008 21:43 To: [EMAIL PROTECTED] Cc: public-lod@w3.org; [EMAIL PROTECTED] Subject: Re: Ordnance Survey data as Linked Data (RE: How do you deprecate URIs? Re: OWL-DL and linked data) Kingsley Idehen wrote: I also forgot to mention obvous use of RDFa in the HTML doc which broadens the range of rdf aware user agents tha commence RDF discovery from HTML Question: is it worth creating a duplicate RDF graph by using RDFa in HTML documents, when there is also RDF/XML available just one link rel=.../ away, and at a distinct URI? Doesn't this RDFa + RDF/XML pattern complicate the RDF-consumption picture in general if we assume agents will want to do something with data aggregated from a number of sources/locations, i.e. doesn't it increase the cost of removing duplicate statements by creating more in the first place? Does it not also complicate the picture of making provenance statements using named graphs, if the subject of the triple could be both an HTML document and an RDF graph? Dunno the answers to these questions, but interested to hear what people think. Tom. -- Tom Heath Researcher Platform Team Talis Information Ltd T: 0870 400 5000 W: http://www.talis.com/platform -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: RDFa + RDF/XML Considered Harmful? (was RE: Ordnance Survey data as Linked Data)
Hugh Glaser wrote: On 14/07/2008 10:42, Mark Birbeck [EMAIL PROTECTED] wrote: [snip] And if we are talking about an RDF browser (as our pages are, albeit with a clean URI that doesn't have the browser URI in it), getting it to include the RDF as RDFa or whatever is even stranger; after all http://demo.openlinksw.com/rdfbrowser2/?uri%5B%5D=http%3A%2F%2Fos.rkbexplorer .com%2Fdescription%2Fosr70017765 doesn't include the substantive RDF as RDFa, (or have a link rel to http://os.rkbexplorer.com/data/osr70017765 for that matter) which would be the equivalent. I can't comment on that example, but ultimately there is no need for a URL for an HTML+RDFa page to be any different to a normal one. (Although I might have missed your point, here) Not sure. I think it relates to the question Tom is asking. Another way of putting it is that we don't expect Tabulator to include RDFa of the RDF we are currently browsing (I think - or maybe we should?). Hugh, So I think it comes down the purpose of the exercise, and there is no one size fits all. If I have some web pages, then being able to simply embed RDF (by whatever means), or link to associated RDF resources, is really great. Good for convenience, maintenance. On the other hand, if I am trying to simply publish RDF, for example out of a DB, then things are a bit different. To help people who want to build agents that use it I might expect them to be able to visualise what I am publishing by using Tabulator or other tools to browse it generically. Or equivalently I might provide some human readable pages of the RDF to make my data more easily accessible. Because it is a bespoke browser, then I need to add rel link to the RDF URI. And then someone suggests I should add RDFa or something more to these pages. Before I know it, I have invested a lot of time providing (and maintaining) neatly formatted and RDF-friendly web pages to publish my DB, when all I wanted to do was join the Semantic Web by publishing my RDF. In fact, one of the options when Kingsley asked for the rel link was to simply remove all the html description pages, getting back to the core of what I want to do, which is publish RDF as Linked Data. Yes, and the little link rel=alternate / addition has extended the scope of your Linked Data deployment to the broader Web :-) And if you choose, you can broaden even further will some RDFa sprinkled in. Note, this is simply an option. Kingsley Best Hugh -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Linked Movie DataBase
Richard Light wrote: In message [EMAIL PROTECTED], Jens Lehmann [EMAIL PROTECTED] writes How do I get it to give me XML RDF? I only seem to be able to get these text triple things, which I don't consider to be a machine-processible format ;-) I get XML when requested. Try: curl -H accept: application/rdf+xml http://data.linkedmdb.org/resource/film/2014 -L Jens, Yes, thanks - so do I. I was trying to get XML RDF by using Firefox with the Tabulator extension, and just expected it to kick in when I went to the data URL. Instead I found myself looking at these text triples. Is there a variant on the URL itself which will return XML RDF? My concern is that XSLT processors should be able to access these linked data resources. XSLT [1.0] only groks XML documents. XSLT just has the document() function in its armoury, and the only information you can give it is a URL: hence the question. (I took this concern to the XSLT list, and the most helpful suggestion I got there was to set up a proxy server which takes URLs, adds an accept header to them, and passes them on. So that's my fallback strategy if I can't put something into the URL to achieve the desired result.) Richard Richard, Please try the Open Data Explorer (esp. Firefox Extension variant [1] ), it works fine with the Linked Movie Database as per some of my twitter micro posts from yesterday [2] . You will also see examples re. other exciting Linked Data additions such as data from the BBC and Southampton Pubs. I even drop a simple mashup in one the links to boot. Links: 1. https://addons.mozilla.org/en-US/firefox/addon/8062 2. http://twitter.com/kidehen -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Visualizing LOD Linkage
Peter Ansell wrote: - Yves Raimond [EMAIL PROTECTED] wrote: Hello! It depends on whether you know that the external references are distinct just based on the URI string. If someone links out to multiple formats using external resource links then they would have to be counted as multiple links as you have no way of knowing that they are different, except in the case where you reserve these types of links as RDF literals. I am not sure if I interpret it correctly - do you mean that you could link to two URIs which are in fact sameAs in the target dataset? Indeed, in that case, the measure would be slightly higher than what it should be. However, I would think that it is rarely (if not never) the case. Peter, I personally don't put sameAs on URI's which relate to the same thing but are really just different representations, ie, the HTML version doesn't get sameAs the RDF version. I really don't believe anyone in this community advocates using owl:sameAs between representations. We use owl:sameAs between Entity URIs while representations are negotiated (content negotiation), discovered via link rel=../, or RDFized etc.. If people knowingly mapped owl:sameAs between representations that weren't identical, then of course this would be flawed if the representations weren't identical. But this isn't what's happening in the LOD space in general, or it's flagship effort: DBpedia. http://dbpedia.org/resource/Berlin is not the representation of the entity Berlin, it's a pointer (Identifier) used by the deploying platform to transmit the description of said entity using a representations desired by the requester/consumer/agent . This entire mechanism isn't new to computing, it's how all our programs work at the lowest levels i.e., we interact with data by reference using pointers [2]. This matter is heart and soul of linked data on the Web or across any other computing medium that manipulates data. Links: 1. http://en.wikipedia.org/wiki/Dereferencable_Uniform_Resource_Identifiers 2. http://cslibrary.stanford.edu/104/ (which clearly needs Linked Data Web variant) Kingsley [SNIP] (link appears to be broken) Springer Link DOI system must be broken. Try the following http://www.springerlink.com/content/w611j82w7v4672r3/fulltext.pdf The following image link is also quite interesting: http://bio2rdf.wiki.sourceforge.net/space/showimage/bio2rdfmap_blanc.png Cheers, Peter -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Fwd: [backstage] Muddy Boots + BBC Music Beta
Chris Sizemore wrote: as i say, the Muddy Boots folks are very LOD-centric and dbpedia-sympathetic, so we should just mention these constructive criticisms to them and i'm sure they'll be taken seriously. also to note is that their work is the result of a BBC commission that i was involved with, so i expect the BBC to make the links between News stories and dbpedia/musicbrainz available itself, eventually. also, the Muddyboots entity extraction source code will be available as open source, i do believe. as rob mentions in his original email, this is proof-of-concept stuff, and linked data wasn't the main use case (though linked data should drop out of it naturally...) Chris, Sure re. these folks, but there are many others that consuming these URIs (as we can see from our logs) . The great thing about URI based Attribution is the fact that you can walk the graphs exposed via the HTTP logs, and check adherence to attribution methods. Lossy attribution value chains remain the biggest hurdle to Linked Data deployment. If data publishers understand the intrinsic power of Linked Data (i.e brand imprint via URIs) combined licensing that requires URI based Attribution, they will be more inclined to publish web documents that expose Linked Data; especially as URI endowed entities distilled from their Web ultimately extend/expand/explode the range of their value propositions. If Attribution by URI isn't unwieldy, then we have a business model unraveling game-changer that extends even beyond the early adapter echelons of LOD :-) Kingsley best-- --cs -Original Message- From: [EMAIL PROTECTED] on behalf of Kingsley Idehen Sent: Wed 8/6/2008 5:19 PM To: Yves Raimond Cc: public-lod@w3.org Subject: Re: Fwd: [backstage] Muddy Boots + BBC Music Beta Yves Raimond wrote: Hello! I thought this would be interesting for this list, as an example of a service using several LOD sources. However, it is a bit sad they don't expose the data they produce as linked data themselves - perhaps we should have a GPL-like license for LOD datasets if you derive data from this linked data, it must be available as linked data :-D (just jocking). Cheers! y -- Forwarded message -- From: robl [EMAIL PROTECTED] Date: Wed, Aug 6, 2008 at 4:15 PM Subject: [backstage] Muddy Boots + BBC Music Beta To: [EMAIL PROTECTED] Hi, Over at Rattle [1] we got quite excited by the release of the new music beta site, in fact we created our own prototype that links our latest iteration of the Muddy Boots system [2] with the beta music site. We haven't announced the latest version of Muddy Boots yet, but in essence it's main aim is to 'unambiguously identify the main actors in a BBC news story', in doing this it uses DBpedia URI's to identify the entities involved. At the moment the system knows about 'people' and companies' however we've just added experimental support for 'bands' (!). As DBpedia knows about Musicbrainz guid's and Muddy Boots knows which BBC news stories relate to which DBpedia entry for a person|band we can add metadata about the related news stories about an artist ('Seen in these news stories' at the bottom of the page) : Coldplay : http://muddy.rattleresearch.com/muddy2/musicentities/cc197bad-dc9c-440d-a5b5-d52ba2e14234 Kaiser Chiefs : http://muddy.rattleresearch.com/muddy2/musicentities/90218af4-4d58-4821-8d41-2ee295ebbe21 and of course not forgetting the famous spoken word artist (!) : George W Bush : http://muddy.rattleresearch.com/muddy2/musicentities/06564917-bdd2-4fb6-bcdc-be9e0c04f7ac We'll be mentioning more about Muddy Boots in the future, but you can see we're starting to add semantic markup to BBC News stories and creating links between open data sources and BBC News (try clicking a news story and you'll see it has the 'actors' in the news story marked up with Microformats). It's still at the prototype stage at the moment and we're about to enter a formal validation and testing phase to measure the systems accuracy - but we thought we'd produce this prototype to demonstrate the kinds of things we can start to achieve when open data (or web-scale) identifiers are used to identify content. You can see all the music entities the system knows about by viewing : http://muddy.rattleresearch.com/muddy2/musicentities/ The support for identifying bands is definitely considered 'experimental' at the moment, so you might see the occasional 'blip' with related stories as we look at how to classify 'bands' in stories more accurately. We're just started indexing BBC news stories in anger, so expect to see more related data appear over the next fews days and weeks. We'd love to hear any comments you have about it :) Thanks, Rob [1] http://www.rattleresearch.com [2] http://muddyboots.rattleresearch.com/semantic-web-project/ - Sent via the backstage.bbc.co.uk discussion group. To unsubscribe
Re: freebase parallax: user interface for browsing graphs of data
Richard H. McCullough wrote: Very nice!!! How is your knowledge base structured? What language? Can I download your program to my computer? Dick McCullough Ayn Rand do speak od mKR done; mKE do enhance od Real Intelligence done; knowledge := man do identify od existent done; knowledge haspart proposition list; http://mKRmKE.org/ David, For purpose of clarity and broader discourse (I know we've been over this in private), what is the Linked Data and/or Semantic Web oriented value of this very cool visualization? Put differently, it would be nice if I could beam a query down the data graph exposed by you very nice visualization rather than being confined to the options presented by your application. At the very least, what's the harm in exposing the Freebase URLs in these Web Pages? If you do that, at the very least, other user agents can do stuff with the graphs (Linked Data) that you are visualizing. The cost of this little tweak is extremely low and the upside extremely high. Cool stuff for sure, but I would like Parallax to be a nice Linked Data Web contribution also :-) Freebase (basic) is part of the Web, but Parallax is sort of confining me to a Freebase enclave by not exposing URLs (where I currently see: javascript:{}). My fundamental argument remains this: Visualizations are good, but graph visualizations are not the sole keys to the treasures that reside within a Graph. Effective traversal (e.g. query beaming SPARQL or MQL) is also part of the puzzle, and it would be nice if we could always offer the visualization and the graph-query-beam-beachhead as part of a single Web information resource deliverable. There are always reasons why one or more humans (due inherent cognitive prowess) would seek to view the same data differently, no matter how compelling the initial visualization, due to the fact that data visualizations are inherently subjective projections. Links: 1. http://tinyurl.com/5of2qu - Abraham Lincoln as Linked Data from Freebase ( I get no triples with Parallax pages since the Freebase Data Sources aren't exposed) Note: The Freebase RDFization Cartridge (Wrapper, Scrapper etc.) will be better i.e., right now there are too many literal property values that should be URIs (this fix is coming). Ditto proper lookup driven meshing with DBpedia. Kingsley - Original Message - From: David Huynh [EMAIL PROTECTED] To: public-lod@w3.org; [EMAIL PROTECTED] Sent: Tuesday, August 12, 2008 3:11 PM Subject: freebase parallax: user interface for browsing graphs of data Hi all, I've been exploring some user interface ideas for browsing graphs (of data in Freebase) in a user-friendly way, which I think might be applicable to some of the data that you have. The screencast should explain: http://mqlx.com/~david/parallax/ Please let me know what you think! Thank you. David -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: freebase parallax: user interface for browsing graphs of data
David Huynh wrote: Kingsley Idehen wrote: David, For purpose of clarity and broader discourse (I know we've been over this in private), what is the Linked Data and/or Semantic Web oriented value of this very cool visualization? Hi Kingsley, David, I apologize for the confusion. I actually didn't say semantic web or web of data or linked open data in my first message. I know, but linkage is inferred by virtue of the forum posts :-) I think Sandro Hawke had a similar response re. congruence. I simply thought that the graph-based browsing paradigm of Parallax might be useful on RDF, which is graph-like. Dbpedia, for example, is a huge graph of billions of triples, and as far as I'm aware, it's hard to explore Dbpedia. I have not doubt about its virtues on the browsing side, none whatsoever. And there has been very similar UI research on browsing graphs, such as Georgi's Humboldt, which I missed out as I wasn't able to attend WWW 2008 :-( Well, at least I had a chance to discuss with him on the Simile mailing list before. But if you would still like to understand the relevance to Linked Data / SW, then may I point you to third parties who have tried to ask/answer that question: http://blogs.zdnet.com/collaboration/?p=131 Excerpt: W3C director Sir Tim Berners-Lee’s long term vision of the Web as a universal medium for data, information, and knowledge exchange would make it possible for the web to understand and satisfy the requests of people and machines. Keyword: Exchange. Key Mechanism: URIs or URLs. All I ask of you is the exposure of the Freebase URLs that are already in sight re. the current / basic Freebase pages. javascript:{} is neither URL nor URI. The ‘people operated’ side of things is the world we live in today. As David Huynh’s video discusses, we busy ourselves manually searching by keyword in multiple locations and then compile our results. Yes, we may be less busy as a result of nice UI (visualization or interaction) but the ultimate goal is stimulation of our inherent cognitive prowess. Thus, also offer users of these UIs an optional route to the raw data. On the Web that's simply about URIs or URLs. A stimulated human is an inquisitive human, and an inquisitive human will ultimately seek to explore dimensions associated with observation, perception, and comprehension. (the talkback comments are also valuable) Can't comment on zdnet as they still require completion of a full subscription interview first, no OpenID or the like in sight :-( http://www.readwriteweb.com/archives/freebase_parallax_taunts_us_wi.php No comment, I think that article speaks for itself. From my vantage point, readwriteweb is still grappling with the fundamentals of what comes after REST APIs. http://blog.futurefacts.net/2008/08/14/future-web-freebase-%E2%80%93-killing-google-with-semantic-precision/ That article is another variant of the readwrite article re. overall comprehension. None of them answered the question re. Linked or Semantic Web relevance, because none of them seem to understand the essence of either. For instance, they don't see how Networking is the constant while points of reference (network entry points) vary over time; we started with Routers (eons ago) and now we are dealing with Entities across Data Spaces [1][2]. Those are their opinions alone. I include them here to provide various perspectives. Here is my basic opinion: I don't have an issue with the UI, it's nice, cool, and innovative. I just want it to be part of the Web. Being part of the Web is about accepting that Linkage is sacrosanct. Thus, no silos, no matter how pretty, cool, or innovative. Your visualization is how I would lead a user to a beachead point in the underlying graph so that they could beam a SPARQL + Full Text pattern query down the graph without them ever writing or seeing a line of SPARQL. Question: Is there any fundamental reason why you cannot expose URIs or URLs where you have javascript:{}? Would this break your work in anyway? If you expose the URIs, I would then be able to demonstrate what I mean using your nice UI, how about that? Links: 1. http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2_TimBL_v3.html#(15) -- Start 2. http://virtuoso.openlinksw.com/presentations/Creating_Deploying_Exploiting_Linked_Data2/Creating_Deploying_Exploiting_Linked_Data2_TimBL_v3.html#(22) -- Where we are headed Note: The presentation above is a remix of presentations by TimBL and I from the recent Linked Data Planet conference. You've forced my hand re. publication as this remix is also about demonstrating RDFa virtues amongst other things re. the Linked Data Web :-) Best, David -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http
Re: freebase parallax: user interface for browsing graphs of data
Georgi Kobilarov wrote: Hi Kingsley, Is this an apropos or an insight addition? My entire position about Linked Data solutions is centered on User Interaction. Note, I said: Visualizations are good, but graph visualizations are not the sole keys to the treasures that reside within a Graph. True. Well, my point was that I got the impression that you've characterized David's interface as a graph visualizer. Which is, imho, not correct. My only issue with Davids work is: javascript:{} where there should be a URI or URL. That's it. And I'd like to highlight the distinction between visualizations and interaction models. The core of every interface is the interaction model. On top of that, there might be specific visualizations. The problem is that most graph UIs I know are based on the one resource at a time interaction model. It's the interaction model of the current web, where a resource is a web page, and users interact by looking at one page (reading it), and browsing from one page to another. That interaction model underlies all current linked data browsers (Tabulator, Disco, etc.). Some UIs use a graph visualization on top, but the interaction model remains the same. (As a side note, I think that visualizing a data as a graph is useless. Why using a 2-dimensional layout where both dimensions are undefined?) But the real magic happens when you change to interaction model, and users can interact with multiple resources at a time. David has shown some excellent examples. Exhibit provided a faceted filtering interaction for graph data. And Parallax now demonstrates a solution to graph browsing. That is part of the magic. There are no panaceas. Another part of the magic is the freedom to flip modes i.e., access to data behind the pretty pages (or other modes of UI). Give users the option to use their own cognitive skills to interact with the data (e.g. pivot, project, and traverse based on their own specific needs). Give users as few options as possible, and design them in a way, that users do not have to use their cognitive skills. Please re-read your comments. In my world view, I don't assume I am smarter than your hypothetical Users. In my world view, I seek to stimulate and arouse the cognitive skills of the users (those who encounter my work). In my world view, I don't assume I have all the answers. In my world view, I believe that harnessing collective intelligence is the ultimate gift of the Web. Cognition is what makes our world tick. It's what guarantees innovation in a timeless continuum. People have to think outside the box or we stagnate and then regress. Programmers cannot acquire the domain expertise of an alien domain. They should. Or at least have an interaction designer supporting them. An excellent programmer and an excellent interaction designer cannot surmount stimulating cognitive power. What you describe is a futile quest (imho). Ironically, you left out systems designers, architects, and data analysts amongst others. Even the sum of all of the missing parts still won't get you close. Kingsley Best, Georgi -- Georgi Kobilarov Freie Universität Berlin www.georgikobilarov.com -Original Message- From: Kingsley Idehen [mailto:[EMAIL PROTECTED] Sent: Thursday, August 14, 2008 2:25 PM To: Georgi Kobilarov Cc: public-lod@w3.org; [EMAIL PROTECTED] Subject: Re: freebase parallax: user interface for browsing graphs of data Georgi Kobilarov wrote: Kingsley, Visualizations are good, but graph visualizations are not the sole keys to the treasures that reside within a Graph. It's not about graph visualization. It's about user interaction for graph-based data. Georgi, Is this an apropos or an insight addition? My entire position about Linked Data solutions is centered on User Interaction. Note, I said: Visualizations are good, but graph visualizations are not the sole keys to the treasures that reside within a Graph. I regard Data Access as part of User Interaction. This is why the OpenLink Data Explorer extension (for instance) simply adds the ability to View | Linked Data Sources to its host browser. All it is really doing is providing the user with a route to a beachead from which a query could be beamed (i.e. SPARQL + Full Text under the covers across the data sources in it's history). Give users the option to use their own cognitive skills to interact with the data (e.g. pivot, project, and traverse based on their own specific needs). We are repeating a long history of not understanding intersection of User Interaction and User Cognitive Skills if we don't provide routes to the data behind subject projections. This matter has been the same since advent of the IT era: 1. Executive Information Systems (what used to be called EIS) 2. Report Writing Systems (Cognos, Crystal Repors, Business Objects
One Last Thing: freebase parallax: user interface for browsing graphs of data
David Huynh wrote: Kingsley Idehen wrote: Question: Is there any fundamental reason why you cannot expose URIs or URLs where you have javascript:{}? Would this break your work in anyway? Just for you, Kingsley, I have fixed it :-) You might need to shift-reload to get the latest code. I wasn't actually expecting such an intense reaction to just javascript:{}. I wonder if that might put off newcomers, who believe that the slightest profanity against The URIs on this mailing list will always trigger such adverse reactions. If you expose the URIs, I would then be able to demonstrate what I mean using your nice UI, how about that? Now that that's all behind us, I'm looking forward to see what you mean. Please, show us what you've got! Best, David David, In the nice to have bucket, is is possible for you to use link rel=dc:source title=Data Sources type=application/ [atom+xml] | [rss+xml] | [rdf+xml] href=feed-information-resource-url / to expose the list of Freebase URLs in you pages. Thus, a list of all URLs (once all javascript:{} have been replaced). Example: http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president Has URLs (instead of: javascript:{}) for each page about a President, but this isn't the case re. Filter Results DIV. Ideally, all the Freebase URLs could be exposed via a Feed which lists all the URLs (as per the link / suggestion above, with dc:source as the DC term for data sources). So one last thing, which cannot be expensive to implement since you are simply substituting a few more javascript:{} for URLs and the listing them is a data sources collection resource (using Atom, RSS 2.0, or RSS 1.0 or even OPML). When all of this is in place, I will then have multiple points from which to launch a SPARQL Query without the user writing a line of SPARQL or seeing anyting to do with RDF. They will simply feel the FORCE of the Linked Data :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: freebase parallax: user interface for browsing graphs of data
Dan Brickley wrote: David Huynh wrote: Kingsley Idehen wrote: Question: Is there any fundamental reason why you cannot expose URIs or URLs where you have javascript:{}? Would this break your work in anyway? Just for you, Kingsley, I have fixed it :-) You might need to shift-reload to get the latest code. I wasn't actually expecting such an intense reaction to just javascript:{}. I wonder if that might put off newcomers, who believe that the slightest profanity against The URIs on this mailing list will always trigger such adverse reactions. Ah, we're an excitable bunch around here ;) Regardless of whether these data URLs are available, the UI work is exciting and I'm glad you shared it with the W3C lists. If others have smart ideas for visualising and navigating RDFesque datasets (whether closed or open by various definitions), let me be clear: they're very welcome to post them to [EMAIL PROTECTED] I hope our enthusiasm hereabouts for open data doesn't discourage people for sharing innovative UI ideas. We need all the help we can get! :) Dan, I hope you understand my response to David was very much in the vein of two-for-one by taking what was a private discussion (between David and I) public for broader knowledge exchange and general discourse purposes. In similar vein, it's well worth SW people poking around http://code.google.com/apis/visualization/ http://processing.org/ and similar, ... the fact that these don't (yet) natively support RDF is no reason not to learn from them... On my part I see User Interaction as the prime issue, with Data Access being part of User Interaction. Once David complete the tweaks I've requested, I believe everything will come together in pretty obvious way. :-) Kingsley cheers, Dan -- http://danbri.org/ -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: One Last Thing: freebase parallax: user interface for browsing graphs of data
David Huynh wrote: Kingsley Idehen wrote: David, In the nice to have bucket, is is possible for you to use link rel=dc:source title=Data Sources type=application/ [atom+xml] | [rss+xml] | [rdf+xml] href=feed-information-resource-url / to expose the list of Freebase URLs in you pages. Thus, a list of all URLs (once all javascript:{} have been replaced). Example: http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president Has URLs (instead of: javascript:{}) for each page about a President, but this isn't the case re. Filter Results DIV. Ideally, all the Freebase URLs could be exposed via a Feed which lists all the URLs (as per the link / suggestion above, with dc:source as the DC term for data sources). So one last thing, which cannot be expensive to implement since you are simply substituting a few more javascript:{} for URLs and the listing them is a data sources collection resource (using Atom, RSS 2.0, or RSS 1.0 or even OPML). When all of this is in place, I will then have multiple points from which to launch a SPARQL Query without the user writing a line of SPARQL or seeing anyting to do with RDF. They will simply feel the FORCE of the Linked Data :-) It's actually not trivial to add that link rel=dc:source / as Parallax is a dynamic mostly client-side application that has only one URL, unless you explicitly ask for permanent links. It's like Google Maps. So, ball is in your court again. :-) There are plenty of proper web links now to work with. Let's see some force :-) It'd be nice to have a screencast of that, too, so it's easy to follow and forward. David David, Okay, will meet you half way for sure. Ball back in my court, and screencast already assumed :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: freebase parallax: user interface for browsing graphs of data
Tim Berners-Lee wrote: On 2008-08 -19, at 21:27, David Huynh wrote: The trouble is of course when the whole web is the database, it's hard to suggest those relationships (connections) for a set of entities. How might one solve that problem? I suppose something like Swoogle can help. Is that what Tabulator uses to know what data is on the SW? Swoogle? Centralized index? No, not at all. The Tabulator is a browser for linked data. The convention for linked data is that if I do a GET on the URI identifying something, the returned information will include the incoming and outgoing links that a reasonable person might be interested in following. Linked data clients dereference any URIs they have not come across before, and pick up either data directly or clues as to where to find what more data. So at each point, you know what the options are leading on. When you pick up a set of related things, of course, there will be some presidents who have children and some who don't. And there will be some presidents which a have bunch of obscure properties. Especially once you have overlaid data from many sources. So then there may be a case for having lenses linked from ontologies to allow one for example to focus on geneology or political career. It gets more complicated when you try to automatically build an interface for allowing people to input data about a president, chose which fields to offer. I'd like to see the Freebase data as linked data on the web ... then we could try all our other UIs on the same data! What would it take to put in some shim interface to Freebase? Tim PS: There are places where a centralized index will help though, like finding people's FOAF page from their email address. David, Which brings us back to the initial point: expose the URIs (you've done this partially) to Linked Data aware agents :-) This, as I've already indicated, comes down to: 1. Replace all javascript:{} with their actual Freebase URIs (which already exist) 2. Use an information resource to collate all the URIs that are in scope at a given point in time in you UI, and then expose the resource URI via link rel=dc:source ../ (you can use RSS, Atom, OPML for this, the key thing is to list the URIs) By doing the above, your Visualization oriented information resource and the data sources behind remain Linked Data Web accessible. Anyway, you've already met me half way (re. yesterday's comments), so I'll crack on with my bit: showing how your interface can provide a beachead for launching a SPARQL Full Text query (without the user writing any SPARQL or seeing any RDF in plain sight). Tim et al: There is a first cut of the last but one Freebase dump in Linked Data form (to the degree attainable) at: http://linkeddata.openlinksw.com:8891/sparql (the graph URI for SPARQL FROM is : http://linkeddata.openlinksw.com/freebase# ) We are working with Francois Belleau (of http://bio2rdf.org) to produce an updated release, but even that release is going to require some work in order for the graph to be totally coherent (*a long story*). Also note, we are RDFizing Freebase on the fly and producing slightly better Linked Data, so if I go to http://mqlx.com/~david/parallax/browse.html?type=%2Fgovernment%2Fus_president and simply do a View | Linked Data Sources via the ODE XPI's context menu against a President's Freebase URI I get: http://tinyurl.com/6pxetc . To conclude, we have Linked Data in two forms for the Freebase Data Space: 1. Their Quad Dumps transformed into RDF and then loaded into Virtuoso resulting in a SPARQL endpoint and a Linked Data Space 2. Linked Data generated on the Fly via our RDFization cartridges with proxy URIs used as mechanism for endowing the resulting Linked Data Space with dereferencable URIs Tim: Tabulator should work with either. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)
Aldo Bucchi wrote: HI, Scanning the thread on Parallax I see some terms reoccurring: Outgoing Connections. Lenses. Lists. Facets. Free text search. IFP to IRI resolution. Find documents that contain IRIs etc... They are all implemented in different ways but tend to share semantics across different browsers and services. How far are we from defining a modular framework so we can mix and math these as atomic interaction pieces? Both services and probably UI parts. Of course, RDF and HTTP can be used to describe them and deliver the descriptions and, in the case of widgets, some OOTB implementations. XForms, Dojo widgets, SWFs? I have done something similar but much simpler in a Flex platform ( I serve Flex modules, described in RDF and referenced by Fresnel vocabs, but only for presentation ). And then on a functional side I have several services that do different things, and I can hot swap them. For example, the free text search service is a (S)WS. Faceter service idem. I guess we still need to see some more diversity to derive a taxonomy and start working on the framework. But it is nice to keep this in sight. The recurring topics. Best, A Aldo, Really nice to see you are looking at things holistically. As you can see, we are veering gradually towards recognizing that the Web, courtesy of HTTP, gives us a really interesting infrasructure for the time-tested MVC pattern (I've been trying to bring attention to this aspect of the Web for a while now). If you look at ODE (closely) you will notice it's an MVC vessel. We have components for Data Access (RDFiztion Cartridges), components for UI (xslt+css templates and fresnel+xslt+css templates), and components for actions (*Cartridges not released yet*). We've tried to focus on the foundation infrastructure that uses HTTP for the messaging across M-V-C so that you get: M--http-V---http---C Unfortunately, our focus on the MC doesn't permeate. Instead, we find all focus coming at us from the V part where we've released minimal templates with hope that 3rd parties will eventually focus on Display Cartridges (via Fresnel, XSLT+SPARQL, xml+xslt+css, etc..). btw - David Schwabe [1] also alluded to the architectural modularity that I am fundamentally trying to bring to broader attention in this evolving conversation re. Linked oriented Web applications. The ultimate goal is to deliver a set of options that enable Web Users to Explore the Web coherently and productively (imho). Humans can only do so much, and likewise Machines, put both together and we fashion a recipe for real collective intelligence (beyond the buzzword). We desperately need to tap into collective intelligenceen route to solving many of the real problems facing the world today. The Web should make seeing and connecting the dots easier, but this is down to the MVC combo as opposed to any single component of the pattern :-) Links: 1. http://lists.w3.org/Archives/Public/public-lod/2008Aug/att-0106/00-part -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Little Correction: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)
[snip] btw - David Schwabe [1] also alluded to the architectural modularity that I am fundamentally trying to bring to broader attention in this evolving conversation re. Linked oriented Web applications. Aldo, In my earlier post, excerpted above, I incorrectly referred to Daniel Schwabe as David Schwabe. This post is my s/David/Daniel style correction, for the record, on our sticky Web :-) The ultimate goal is to deliver a set of options that enable Web Users to Explore the Web coherently and productively (imho). Humans can only do so much, and likewise Machines, put both together and we fashion a recipe for real collective intelligence (beyond the buzzword). We desperately need to tap into collective intelligenceen route to solving many of the real problems facing the world today. The Web should make seeing and connecting the dots easier, but this is down to the MVC combo as opposed to any single component of the pattern :-) Links: 1. http://lists.w3.org/Archives/Public/public-lod/2008Aug/att-0106/00-part -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: soliciting your favorite (public) SPARQL queries!
Lee Feigenbaum wrote: Hi everyone, I'm putting together a SPARQL by Example tutorial, which is, as the name suggests, a step-by-step introduction to SPARQL taught almost entirely through complete, runnable SPARQL queries. So far, I've gathered a great deal of example queries myself, but I know that many subscribers to these lists probably have favorite queries of their own that you might be willing to share with me. I'm looking for: 1) SPARQL queries 2) ...that can be run by anyone (no private data sets) 3a)...either by running the query against a public SPARQL endpoint 3b)...or by using a public SPARQL endpoint that will fetch HTTP-accessible RDF data (e.g. sparql.org or demo.openlinksw.com) 4) ...that answers a real* question 5) ...and that is fun!** * real is in the eye of the beholder, I imagine, but I'm not looking for finds the predicates that relate ex:s and ex:o in this sample RDF graph ** fun is also in the eye of the beholder. fun can be a query on fun data; a clever query that may illustrate a particular SPARQL construct (trick); a query that integrates interesting information; a query with surprising results; etc. thanks to anyone who is able to contribute! Lee PS I plan to make the tutorial slides available online under an appropriate CC license once they are completed. Lee, Here are our contributions (endpoint: http://demo.openlinksw.com/sparql): 1. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/NorthWindREF - against live RDF Views of Northwind SQL Database (Products, Customers, Orders etc.) 2. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSSIOCRef - collection of SPARQL queries against SIOC based exploration of OpenLink Data Space s blogs, wikis, bookmarks, calendars, feed subscriptions etc..) 3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSFOAFRef - collection of SPARQL queries for FOAF navigation of OpenLink Data Spaces 4. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSSKOSRef - collection of SPARQL queries for SKOS based exploration of Tag Spaces associated with OpenLink Data Spaces 5. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/SIOCRefDocs- collection of SPARQL queries using SIOC to explore Virtuoso's Online Docs Data Space 6. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/WordPressSIOCRef - SPARQL queries for exploring Wordpress Data when Deployed via Virtuoso's PHP runtime hosting 7. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/MediaWikiSIOCRef - ditto re. MediaWiki 8. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/PHPBB3SIOCRef - ditto phpBB 9. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/DrupalSIOCRef - ditto Drupal 10. http://www4.wiwiss.fu-berlin.de/benchmarks-200801/#queries - DBpedia Benchmark Queries (* run against the DBpedia SPARQL endpoint at: http://dbpedia.org/sparql *) 11. http://tinyurl.com/5fjp5h - Collection of Inference based Queries against DBpedia that leverage the Yago Class Hierarchy Note: re. the ODS examples (items 2 - 4 ), if you change the SPARQL endpoint to: http://community.linkeddata.org/sparql , and Graph IRI in the SPARQL FROM clause to: http://community.linkeddata.org/sparql, you can then run the SIOC, FOAF, SKOS examples against the live LOD Data Space with more interesting results. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Modular Browsers (was RE: freebase parallax: user interface for browsing graphs of data)
Aldo Bucchi wrote: Hello, ( replies inlined ) On Wed, Aug 20, 2008 at 9:19 PM, Kingsley Idehen [EMAIL PROTECTED] wrote: Aldo Bucchi wrote: HI, Scanning the thread on Parallax I see some terms reoccurring: Outgoing Connections. Lenses. Lists. Facets. Free text search. IFP to IRI resolution. Find documents that contain IRIs etc... They are all implemented in different ways but tend to share semantics across different browsers and services. How far are we from defining a modular framework so we can mix and math these as atomic interaction pieces? Both services and probably UI parts. Of course, RDF and HTTP can be used to describe them and deliver the descriptions and, in the case of widgets, some OOTB implementations. XForms, Dojo widgets, SWFs? I have done something similar but much simpler in a Flex platform ( I serve Flex modules, described in RDF and referenced by Fresnel vocabs, but only for presentation ). And then on a functional side I have several services that do different things, and I can hot swap them. For example, the free text search service is a (S)WS. Faceter service idem. I guess we still need to see some more diversity to derive a taxonomy and start working on the framework. But it is nice to keep this in sight. The recurring topics. Best, A Aldo, Really nice to see you are looking at things holistically. I showed up with a narrow interface, I know ;) As you can see, we are veering gradually towards recognizing that the Web, courtesy of HTTP, gives us a really interesting infrasructure for the time-tested MVC pattern (I've been trying to bring attention to this aspect of the Web for a while now). If you look at ODE (closely) you will notice it's an MVC vessel. We have components for Data Access (RDFiztion Cartridges), components for UI (xslt+css templates and fresnel+xslt+css templates), and components for actions (*Cartridges not released yet*). Ah... I remember telling Daniel Lewis something was missing from his UPnP diagram: a way to modify the Model. aka: a Controller / Actions. You are right, technically an agent like ODE ( assuming you can hook in actions ) is all that you need to allow users to interact with linked data. Let's say that this sort of solution can cover 80% of user interaction cases ( launching simple actions and direct manipulation of resources ), and operates on top of 80% of data ( anything that can be published as linked data/SPARQL and fits within the expressiveness of RDF's abstract model ). Not a bad MVC structure at all! So, how do you plan on hooking up the actions to the shell, is this in the cartridges? How will they surface. Context menu? Everything lives in a REST or SOAP accessible Cartridge. ODE just talks REST or SOAP . For instance, ODE uses REST calls to the Sponger Service when RDFizing, but it can just as well use Triplr. We've just put out a new ODE release with an improved Preferences dialog that makes the mixing and matching of Renderers and RDFizers clearer. Re. Display Cartridges (Fresnel Templates) the same would apply, but in this case we just deal with the URIs of the templates. Ditto in the case of Actions. URIs and REST are all that we need fundamentally, which is just about leveraging what the Web has always offered. We've tried to focus on the foundation infrastructure that uses HTTP for the messaging across M-V-C so that you get: M--http-V---http---C Unfortunately, our focus on the MC doesn't permeate. Instead, we find all focus coming at us from the V part where we've released minimal templates with hope that 3rd parties will eventually focus on Display Cartridges (via Fresnel, XSLT+SPARQL, xml+xslt+css, etc..). Well. The M part is the data, isn't it? ( so it is permeating, people are publishing data ). Yes. Unless you mean building some higher functionality services ( on top of SPARQL and RDF ) such as faceters, free text search, IFP resolution, etc. But in that case it is also moving forward, although not with a standardized interface. This could be thought of as higher level Data Access components. Correct, with Linked Data at the base. This is why I always refere to Linked Data as the foundation layer of the Semantic Web innovation continuum. The C part... that's another story. As I pointed out before, you need to define the way and an environment to hook in the actions. What is the shell? ODE is an inner shell / enclave example, but in reality the Web is the outer shell (as long as you honor links and link dereferencing). What is missing is a vocabulary for actions. Example: if you combined GoodRelations Ontology and Web Services part of SIOC (the extended modules part) you get a Linked Data Space that not only describes a service vendor, but also one that formally describes how to consummate a transaction with said vendor, with granularity at the service level (so two services with different signatures from
Re: soliciting your favorite (public) SPARQL queries!
and ex:o in this sample RDF graph ** fun is also in the eye of the beholder. fun can be a query on fun data; a clever query that may illustrate a particular SPARQL construct (trick); a query that integrates interesting information; a query with surprising results; etc. thanks to anyone who is able to contribute! Lee PS I plan to make the tutorial slides available online under an appropriate CC license once they are completed. -- Dr. Axel Polleres, Digital Enterprise Research Institute (DERI) email: [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] url: http://www.polleres.net/ Everything is possible: rdfs:subClassOf rdfs:subPropertyOf rdfs:Resource. rdfs:subClassOf rdfs:subPropertyOf rdfs:subPropertyOf. rdf:type rdfs:subPropertyOf rdfs:subClassOf. rdfs:subClassOf rdf:type owl:SymmetricProperty. Adrian, Re. SPARQL Aggregates, see: http://esw.w3.org/topic/SPARQL/Extensions/Aggregates -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Mozilla Labs Ubiquity and Linked Data
Hausenblas, Michael wrote: UPDATE: After a rather short night (I tell you, toying around with ubiquity is *really* addictive) there are now two more commands available: one for querying DBpedia (simple skos:subject on categories) and a curl HEAD command (should be self-explanatory). If you develop (Web of Data) commands as well, please consider announcing them on the Mozilla Wiki [1]. Cheers, Michael [1] https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH http://www.joanneum.at/iis/ -- Here is our initial Ubiquity command for Linked Data which has the command pattern: describe-page URL: http://demo.openlinksw.com/ubiq/ Of course there is more to come re. a generic SPARQL command etc.. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Correction: Mozilla Labs Ubiquity and Linked Data
Kingsley Idehen wrote: Hausenblas, Michael wrote: UPDATE: After a rather short night (I tell you, toying around with ubiquity is *really* addictive) there are now two more commands available: one for querying DBpedia (simple skos:subject on categories) and a curl HEAD command (should be self-explanatory). If you develop (Web of Data) commands as well, please consider announcing them on the Mozilla Wiki [1]. Cheers, Michael [1] https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH http://www.joanneum.at/iis/ -- Here is our initial Ubiquity command for Linked Data which has the command pattern: describe-page URL: http://demo.openlinksw.com/ubiq/ Command is currently: page-describe URL (as per index.html), but it will ultimately be: describe-page URL :-) Kingsley Of course there is more to come re. a generic SPARQL command etc.. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Mozilla Labs Ubiquity and Linked Data
Hausenblas, Michael wrote: Kingsley, Great work! Again, OpenLink shows that it is not only a good but also a very quick player in the linked data field. If I only had the time now (/me currently at TRIPLE-I conference and busy applauding Tom ;) I'd have done it myself, sorry. Anyways, good to see the uptake. Two minor comments: (i) the command, I guess actually is 'page-describe' and (ii) we might wanna make it configurable to play with other SW browsers such as Tabulator (as you proposed, IIRC). Michael, I think it does play well with the other Linked Data clients (see: page footer) :-) Kingsley Cheers, Michael -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH Steyrergasse 17, A-8010 Graz, AUSTRIA office phone: +43-316-876-1193 (fax:-1191) e-mail: [EMAIL PROTECTED] web: http://www.joanneum.at/iis/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.joanneum.at/iis/ private mobile: +43-660-7621761 web: http://www.sw-app.org/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.sw-app.org/ -- From: [EMAIL PROTECTED] on behalf of Kingsley Idehen Sent: Thu 2008-09-04 17:20 To: public-lod@w3.org Cc: [EMAIL PROTECTED]; public-lod@w3.org Subject: Re: Mozilla Labs Ubiquity and Linked Data Hausenblas, Michael wrote: UPDATE: After a rather short night (I tell you, toying around with ubiquity is *really* addictive) there are now two more commands available: one for querying DBpedia (simple skos:subject on categories) and a curl HEAD command (should be self-explanatory). If you develop (Web of Data) commands as well, please consider announcing them on the Mozilla Wiki [1]. Cheers, Michael [1] https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH http://www.joanneum.at/iis/ -- Here is our initial Ubiquity command for Linked Data which has the command pattern: describe-page URL: http://demo.openlinksw.com/ubiq/ Of course there is more to come re. a generic SPARQL command etc.. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com http://www.openlinksw.com/ -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Mozilla Labs Ubiquity and Linked Data Other Linked Data Clients
Hausenblas, Michael wrote: Kingsley, Great work! Again, OpenLink shows that it is not only a good but also a very quick player in the linked data field. If I only had the time now (/me currently at TRIPLE-I conference and busy applauding Tom ;) I'd have done it myself, sorry. Anyways, good to see the uptake. Two minor comments: (i) the command, I guess actually is 'page-describe' and (ii) we might wanna make it configurable to play with other SW browsers such as Tabulator (as you proposed, IIRC). Cheers, Michael -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH Steyrergasse 17, A-8010 Graz, AUSTRIA office phone: +43-316-876-1193 (fax:-1191) e-mail: [EMAIL PROTECTED] web: http://www.joanneum.at/iis/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.joanneum.at/iis/ private mobile: +43-660-7621761 web: http://www.sw-app.org/ https://webmail.joanneum.at/exchweb/bin/redir.asp?URL=http://www.sw-app.org/ -- From: [EMAIL PROTECTED] on behalf of Kingsley Idehen Sent: Thu 2008-09-04 17:20 To: public-lod@w3.org Cc: [EMAIL PROTECTED]; public-lod@w3.org Subject: Re: Mozilla Labs Ubiquity and Linked Data Hausenblas, Michael wrote: UPDATE: After a rather short night (I tell you, toying around with ubiquity is *really* addictive) there are now two more commands available: one for querying DBpedia (simple skos:subject on categories) and a curl HEAD command (should be self-explanatory). If you develop (Web of Data) commands as well, please consider announcing them on the Mozilla Wiki [1]. Cheers, Michael [1] https://wiki.mozilla.org/Labs/Ubiquity/Commands_In_The_Wild#Web_Of_Data -- Michael Hausenblas, MSc. Institute of Information Systems Information Management JOANNEUM RESEARCH Forschungsgesellschaft mbH http://www.joanneum.at/iis/ -- Here is our initial Ubiquity command for Linked Data which has the command pattern: describe-page URL: http://demo.openlinksw.com/ubiq/ Of course there is more to come re. a generic SPARQL command etc.. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com http://www.openlinksw.com/ Michael, I just realized that you may have been requesting that we use our Proxy URIs in the calls we make to the other Linked Data Clients (in the page footer), which is how we feed them units of RDFization via our Proxy/Wrapper URIs. That's certainly worth doing. There will be an update with this feature. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Announce: OpenLink Data Explorer Extension Update
All, We are please to announce the availability of an update release of the OpenLink Data Explorer (ODE) extension for Firefox [1][2]. Features and enhancements include: 1. New Page Description main and context menu item that provides a an (X)HTML representation of a resource description 2. First look at new Find feature which replaces Search and seeks to demonstrate that Linked Data helps us Find rather than Search for stuff (more to come re. this powerful feature in follow-on releases) 3. A number of aesthetic fixes We've also updated the ubiquity commands [3] that provide commandline access points to the same services that drive ODE. Links: 1. https://addons.mozilla.org/en-US/firefox/addon/8062 2. http://ode.openlinksw.com/#Download 3. http://demo.openlinksw.com/ubiq (*read my most recent blog post re. subscribe and unsubscribe howto*) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
World Bank API
All, For those unaware of this development, the World Bank has now released APIs for their data space on the Web [1]. Links: 1. http://developer.worldbank.org/ -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: New Calais proxy could grow Linked Data Cloud
Paul Miller wrote: Members of this list might be interested in my write-up of ThomsonReuters' latest beta service... which I think will prove pretty useful in growing the Linked Data cloud... especially for news content from the BBC et al... http://blogs.zdnet.com/semantic-web/?p=194 Paul -- Paul Miller Technology Evangelist, Talis w: www.talis.com/ http://www.talis.com/ skype: napm1971 mobile/cell: +44 7769 740083 http://blogs.zdnet.com/semantic-web/ _www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er_ Paul, How does this actually benefit or contribute to the Linked Data Cloud? I ask specifically because URIs (of the dereferencable variety) are missing in action. Hopefully, I am completely overlooking something here :-) We tend to use the term Proxy or Wrapper to describe solutions in the Linked Data realm that generate dereferencable URIs based RDF graphs (aka. Linked Data Spaces) on the fly via RDF-ization middleware. If possible, please encourage the OpenCalais folks (Tom et al.) to respond to my comments above via a response to this post. Example Proxy / Wrapper URIs in the wild: 1. http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/abraham_lincoln - Document about Abraham Lincoln 2. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.freebase.com/view/en/abraham_lincoln%23this - Abraham Lincoln the Entity of type foaf:Person that is also a sioc:Item 3. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.crunchbase.com/company/thomson-reuters%23this - Thompson Reuters the Entity of type foaf:Organization that is also a sioc:Item 4. http://www4.wiwiss.fu-berlin.de/flickrwrappr/photos/Thomson_Reuters - Thompson Reuters photos from Flickr 5. http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fwww4.wiwiss.fu-berlin.de%2Fflickrwrappr%2Fphotos%2FThomson_Reuters - Browser view of the data space exposed by the Flickr wrapper URI -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Freebase wrapper
Georgi Kobilarov wrote: Hi Kingsley, Example Proxy / Wrapper URIs in the wild: 1. http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/a braham_lincoln - Document about Abraham Lincoln 2. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/ rdf/http://www.freebase.com/view/en/abraham_lincoln%23this - Abraham Lincoln the Entity of type foaf:Person that is also a sioc:Item these URIs unfortunately do not show any data. Would be great to see your freebase wrapper in action... Best, Georgi Georgi, As per usual, I announce a demo at a time when maintenance occurs on the demo server :-( See: http://demo.openlinksw.com/about/html/http://www.freebase.com/view/en/abraham_lincoln Kingsley -- Georgi Kobilarov Freie Universität Berlin www.georgikobilarov.com -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Kingsley Idehen Sent: Tuesday, September 23, 2008 2:02 PM To: Paul Miller Cc: public-lod@w3.org Subject: Re: New Calais proxy could grow Linked Data Cloud Paul Miller wrote: Members of this list might be interested in my write-up of ThomsonReuters' latest beta service... which I think will prove pretty useful in growing the Linked Data cloud... especially for news content from the BBC et al... http://blogs.zdnet.com/semantic-web/?p=194 Paul -- Paul Miller Technology Evangelist, Talis w: www.talis.com/ http://www.talis.com/ skype: napm1971 mobile/cell: +44 7769 740083 http://blogs.zdnet.com/semantic-web/ _www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er_ Paul, How does this actually benefit or contribute to the Linked Data Cloud? I ask specifically because URIs (of the dereferencable variety) are missing in action. Hopefully, I am completely overlooking something here :-) We tend to use the term Proxy or Wrapper to describe solutions in the Linked Data realm that generate dereferencable URIs based RDF graphs (aka. Linked Data Spaces) on the fly via RDF-ization middleware. If possible, please encourage the OpenCalais folks (Tom et al.) to respond to my comments above via a response to this post. Example Proxy / Wrapper URIs in the wild: 1. http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/a braham_lincoln - Document about Abraham Lincoln 2. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/ rdf/http://www.freebase.com/view/en/abraham_lincoln%23this - Abraham Lincoln the Entity of type foaf:Person that is also a sioc:Item 3. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/ rdf/http://www.crunchbase.com/company/thomson-reuters%23this - Thompson Reuters the Entity of type foaf:Organization that is also a sioc:Item 4. http://www4.wiwiss.fu-berlin.de/flickrwrappr/photos/Thomson_Reuters - Thompson Reuters photos from Flickr 5. http://demo.openlinksw.com/rdfbrowser2/?uri=http%3A%2F%2Fwww4.wiwiss.fu -berlin.de%2Fflickrwrappr%2Fphotos%2FThomson_Reuters - Browser view of the data space exposed by the Flickr wrapper URI -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: New Calais proxy could grow Linked Data Cloud
Thomas Tague wrote: LOD Group: Tom, First off, welcome! Really happy to see you've taken the time engage the LOD community re. your team's efforts and long term objectives. Questions follow inline below. First, a philosophical point and then a few facts. When your child first learns to read you don't discard that because they haven't yet graduated from college. You know college is coming, you're already thinking about college, you may actually be actively working on college - but the first words are still important. Calais is learning to read. We firmly believe in releasing building blocks when they become available rather than waiting (and waiting and waiting) for the entire solution to be ready. A few specific facts to make it clearer where SemanticProxy fits in: 1) We will have de-referenceable URIs for every entity extracted by Calais by the end of this year. The engineering is done and we're in active design and build mode. We haven't finished the analysis yet - but this will be millions of endpoints on the day we go live. Please clarify what you mean by endpoints. Over here it might refer to SPARQL endpoints or derferencable URIs. I suspect you mean URIs, but clarification from you will aid others. 2) A *subset* of those entity types will absolutely have links to other linked data sources when we go live. Right now we know there will be substantive links for companies, geographies and a few of the easy ones like music, books, etc. We'll expand on that set over time and have a goal of setting up a community-based mechanism for enhancing the links over time. Will linkage apply to instance data and associated definitions data (ontology / schema / data dictionary) for the Thompson Reuters linked data spaces? Will you be using shared ontologies where such exist, or at the very least put out your ontology in RDFS or OWL? Even doing this open up the doors for community participation in the data definitions linkage effort (e.g. what's happened re. UMBEL, OpenCyc, Yago, and Wordnet). 3) At the end of this month (September) as part of Release 3.1 we'll be releasing company and geography disambiguation as a component of the metadata generation process. The company disambiguation is based on a lexicon of over 16M company aliases + additional hinting and we have a similar approach with geography. Great news, but the real utility of such work will always be easier to imbibe, by this community in particular, if the resulting output is an RDF Linked Data Space rather than an RDF Data Island :-) Again, great to have you outline your development road-map here, I certainly believe this will ultimately be a great contribution to the burgeoning Linked Data Web. Kingsley Question? Ideas? Fire away. Tom On Tue, Sep 23, 2008 at 8:21 AM, Paul Miller [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote: From the post... SemanticProxy will return dereferenceable Linked Data URIs by the end of this quarter. Paul -- Paul Miller Technology Evangelist, Talis w: www.talis.com/ http://www.talis.com/ skype: napm1971 mobile/cell: +44 7769 740083 http://blogs.zdnet.com/semantic-web/ _www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er_ On 23 Sep 2008, at 13:02, Kingsley Idehen wrote: Paul Miller wrote: Members of this list might be interested in my write-up of ThomsonReuters' latest beta service... which I think will prove pretty useful in growing the Linked Data cloud... especially for news content from the BBC et al... http://blogs.zdnet.com/semantic-web/?p=194 Paul -- Paul Miller Technology Evangelist, Talis w: www.talis.com/ http://www.talis.com/ http://www.talis.com/ skype: napm1971 mobile/cell: +44 7769 740083 http://blogs.zdnet.com/semantic-web/ _www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er http://www.linkedin.com/in/pau1mi11er_ Paul, How does this actually benefit or contribute to the Linked Data Cloud? I ask specifically because URIs (of the dereferencable variety) are missing in action. Hopefully, I am completely overlooking something here :-) We tend to use the term Proxy or Wrapper to describe solutions in the Linked Data realm that generate dereferencable URIs based RDF graphs (aka. Linked Data Spaces) on the fly via RDF-ization middleware. If possible, please encourage the OpenCalais folks (Tom et al.) to respond to my comments above via a response to this post. Example Proxy / Wrapper URIs in the wild: 1. http://demo.openlinksw.com/proxy/html/http://www.freebase.com/view/en/abraham_lincoln - Document about Abraham Lincoln 2. http://demo.openlinksw.com/about/html/http://demo.openlinksw.com/about/rdf/http://www.freebase.com/view/en/abraham_lincoln%23this - Abraham Lincoln the Entity of type
Re: New Calais proxy could grow Linked Data Cloud
Tom Heath wrote: Hi Chris, 2008/9/23 Chris Sizemore [EMAIL PROTECTED]: does anyone else expect/hope that the Linked Data part of dereferenceable Linked Data URIs implies that these URIs will be linked up to, for instance, dbPedia, umbel, musicbrainz, etc etc? otherwise it wouldn't really be linked data per se, no? I hope that that's the case, although I don't know if it will turn out that way. It may be a case of Calais confusing Linked Data with linkable data. It is certainly my expectation (and that of the entire LD community, I would argue) that Linked Data does actually include those links, in case anyone was in any doubt ;) Tom. Chris / Tom, I would decompose Linked Data, in this context, as follows: 1. URIs can be used to de-reference an RDF based resource description 2. Resource description includes de-rferencable URIs across Subject, Predicate, and Object (where applicable) in constituent triples Linkage to other Linked Data spaces (instances and classes) is a desired added bonus, but not the essential definition (imho) :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL
Paul Gearon wrote: On Mon, Sep 22, 2008 at 3:47 AM, Eyal Oren [EMAIL PROTECTED] wrote: On 09/19/08/09/08 23:12 +0200, Orri Erling wrote: Has has there been any analysis on whether there is a *fundamental* reason for such performance difference? Or is it simply a question of maturity; in other words, relational db technology has been around for a very long time and is very mature, whereas RDF implementations are still quite recent, so this gap will surely narrow ...? This is a very complex subject. I will offer some analysis below, but this I fear will only raise further questions. This is not the end of the road, far from it. As far as I understand, another issue is relevant: this benchmark is somewhat unfair as the relational stores have one advantage compared to the native triple stores: the relational data structure is fixed (Products, Producers, Reviews, etc with given columns), while the triple representation is generic (arbitrary s,p,o). This point has an effect on several levels. For instance, the flexibility afforded by triples means that objects stored in this structure require processing just to piece it all together, whereas the RDBMS has already encoded the structure into the table. Ironically, this is exactly the reason we (Tucana/Kowari/Mulgara) ended up building an RDF database instead of building on top of an RDBMS: The flexibility in table structure was less efficient that a system that just knew it only had to deal with 3 columns. Obviously the shape of the data (among other things) dictates what it is the better type of storage to use. A related point is that processing RDF to create an object means you have to move around a lot in the graph. This could mean a lot of seeking on disk, while an RDBMS will usually find the entire object in one place on the disk. And seeks kill performance. This leads to the operations used to build objects from an RDF store. A single object often requires the traversal of several statements, where the object of one statement becomes the subject of the next. Since the tables are typically represented as Subject/Predicate/Object, this means that the main table will be joined against itself. Even RDBMSs are notorious for not doing this efficiently. One of the problems with self-joins is that efficient operations like merge-joins (when they can be identified) will still result in lots of seeking, since simple iteration on both sides of the join means seeking around in the same data. Of course, there ARE ways to optimize some of this, but the various stores are only just starting to get to these optimizations now. Relational databases suffer similar problems, but joins are usually only required for complex structures between different tables, which can be stored on different spindles. Contrast this to RDF, which needs to do do many of these joins for all but the simplest of data. One can question whether such flexibility is relevant in practice, and if so, one may try to extract such structured patterns from data on-the-fly. Still, it's important to note that we're comparing somewhat different things here between the relational and the triple representation of the benchmark. This is why I think it is very important to consider the type of data being stored before choosing the type of storage to use. For some applications an RDBMS is going to win hands down every time. For other applications, an RDF store is definitely the way to go. Understanding the flexibility and performance constraints of each is important. This kind of benchmarking helps with that. It also helps identify where RDF databases need to pick up their act. Regards, Paul Gearon Paul, You make valid points, the problem here is that the benchmark has been released without enough clarity about it's prime purpose. To even compare RDF Quads Stores with an RDBMS engine when the schema is Relational in itself is kinda twisted. The role of mappers (DR2Q Virtuoso RDF Views) for instance, should have been made much clearer, maybe in separate results tables. I say this because these mappers offer different approaches to projecting RDBMS based data in RDF Linked Data form, on the fly, and their purpose in this benchmark is all about raw performance and scalability as it relates to following RDF Linked Data generation and deployment conditions: 1. Schema is Relational 2. RDF warehouse is impractical As I am sure you know, we could invert this whole benchmark Open World style, and then bring RDBMS engines to their knees by incorporating SPARQL query patterns comprised of ?p's and subclasses . To conclude, the quad store numbers should simply be a conparisons of the quad stores themselves, and not the quad stores vs the mappers or native SQL. This clarification really needs to make it's way into the benchmark narrative. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO
Re: AW: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL
Chris Bizer wrote: Hi Kingsley and Paul, Yes, I completely agree with you that different storage solutions fit different use cases and that one of the main strengths of the RDF data model is its flexibility and the possibility to mix different schemata. Nevertheless, it think it is useful to give application developers an indicator about what performance they can expect when they choose a specific architecture, which is what the benchmark is trying to do. Chris, Yes, but the user profile has to be a little clearer. If you separate the results in the narrative you achieve the goal. You can use SQL numbers as a sort of benchamark if you clearly explain the nature skew that SQL enjoys due to the nature of the schema. We plan to run the benchmark again in January and it would be great to also test Tucana/Kowari/Mulgara in this run. As the performance of RDF stores is constantly improving, let's also hope that the picture will not look that bad for them anymore then. But at the current time, there is no clear sense of what better means :-) What's the goal? What I fundamentally take from the benchmarks are the following: 1. Native RDF and RDF Views/Mapper scalability is becoming less of an issue (of course depending on your choice of product) and we are already at the point where this technology can be used for real-world solutions that have enterprise level scalability demands and expectations 2. It's impractical to create RDF warehouses from a existing SQL Data Sources when you can put RDF Views / Wrappers in front of the SQL Data Sources (SQL cost optimization technology has evolved significantly over the years across RDBMS engines). And Yes, I would also like to see Mulgara and others RDF Stores in the next round of benchmarks :-) Kingsley Cheers, Chris -Ursprüngliche Nachricht- Von: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Im Auftrag von Kingsley Idehen Gesendet: Mittwoch, 24. September 2008 20:57 An: Paul Gearon Cc: [EMAIL PROTECTED]; public-lod@w3.org Betreff: Re: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL Paul Gearon wrote: On Mon, Sep 22, 2008 at 3:47 AM, Eyal Oren [EMAIL PROTECTED] wrote: On 09/19/08/09/08 23:12 +0200, Orri Erling wrote: Has has there been any analysis on whether there is a *fundamental* reason for such performance difference? Or is it simply a question of maturity; in other words, relational db technology has been around for a very long time and is very mature, whereas RDF implementations are still quite recent, so this gap will surely narrow ...? This is a very complex subject. I will offer some analysis below, but this I fear will only raise further questions. This is not the end of the road, far from it. As far as I understand, another issue is relevant: this benchmark is somewhat unfair as the relational stores have one advantage compared to the native triple stores: the relational data structure is fixed (Products, Producers, Reviews, etc with given columns), while the triple representation is generic (arbitrary s,p,o). This point has an effect on several levels. For instance, the flexibility afforded by triples means that objects stored in this structure require processing just to piece it all together, whereas the RDBMS has already encoded the structure into the table. Ironically, this is exactly the reason we (Tucana/Kowari/Mulgara) ended up building an RDF database instead of building on top of an RDBMS: The flexibility in table structure was less efficient that a system that just knew it only had to deal with 3 columns. Obviously the shape of the data (among other things) dictates what it is the better type of storage to use. A related point is that processing RDF to create an object means you have to move around a lot in the graph. This could mean a lot of seeking on disk, while an RDBMS will usually find the entire object in one place on the disk. And seeks kill performance. This leads to the operations used to build objects from an RDF store. A single object often requires the traversal of several statements, where the object of one statement becomes the subject of the next. Since the tables are typically represented as Subject/Predicate/Object, this means that the main table will be joined against itself. Even RDBMSs are notorious for not doing this efficiently. One of the problems with self-joins is that efficient operations like merge-joins (when they can be identified) will still result in lots of seeking, since simple iteration on both sides of the join means seeking around in the same data. Of course, there ARE ways to optimize some of this, but the various stores are only just starting to get to these optimizations now. Relational databases suffer similar problems, but joins are usually only required for complex
Re: Announcing Open GUID
[EMAIL PROTECTED] wrote: Yes, I have not loaded DBPedia yet. I started with WordNet 3.0, which apparently did not deem IBM as a part of our language (note Mircosoft and Google do make the cut!). See the (http://openguid.net/roadmap) for more details on the data loading initiative. Once that is complete, if your favorite (least favorite?) company is still not loaded, anyone will be able to add it, tag it, and add identical references. And yes, the function was to find you an existing Open GUID based on some text. I will try to make that clearer. Thanks for the test drive, Jason Jason, Yes, I see some data now [1]. How are you planning to proceed re. data from UMBEL, Yago, OpenCyc, and DBpedia? Links: 1. http://demo.openlinksw.com/proxy/html/http://openguid.net/e6adf4e8-da25-102b-9a03-2db401e887ec Kingsley Original Message Subject: Re: Announcing Open GUID From: Kingsley Idehen [EMAIL PROTECTED] Date: Thu, September 25, 2008 8:48 am To: Damian Steer [EMAIL PROTECTED] Cc: [EMAIL PROTECTED], public-lod@w3.org Damian Steer wrote: -BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Kingsley Idehen wrote: | I am exploring your service. | | Please clarify this: | http://openguid.net/search?keyword=http%3A%2F%2Fdbpedia.org%2Fresource%2FIBM | | | The demo asked me to input an Open GUID (a URI) re. the Find feature. Do you mean 'Find GUID'? I read that as meaning 'Find me a GUID for this text I've just entered'. Damian, Yes. Okay, I tried: http://openguid.net/search?keyword=ibm Didn't get a URI back. Sindice give me: http://sindice.com/search?q=ibmqt=term Maybe the DBpedia binding hasn't take place yet? I am hoping to get the URIs associated with IBM the entity of Type some-ontology:some-term-example-organization Kingsley Damian -BEGIN PGP SIGNATURE- Version: GnuPG v1.4.6 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org iD8DBQFI25r3AyLCB+mTtykRAm50AKCMvpRUjv0XOabY7Q9Kcg9kuAV/7gCfZj2t o1plwBi8zNimV+UGHmxAL2k= =G1Ix -END PGP SIGNATURE- -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Drilling into the LOD Cloud
Dan Brickley wrote: Kingsley Idehen wrote: Ed Summers wrote: On Mon, Sep 22, 2008 at 1:28 PM, Giovanni Tummarello [EMAIL PROTECTED] wrote: I really support your idea for a lod cloud that's actually useful to write queries, i promise we'll do the best from sindice to deliver one such a thing. That sounds like a great idea for a service. Right now I just need a little diagram that illustrates the links between resources in the LOD cloud; and the diversity of descriptions from each provider. But, I guess I should stop begging on here, and just create it eh? Thanks for the feedback, //Ed Ed, Are you not able to add additional rdf:type links between DBpedia and the following: 1. Yago 2. OpenCyc 3. UMBEL Then between UMBEL and OpenCyc: 1. owl:sameAs 2. owl:equivalentClass If these thingies are owl:sameAs, then presumably they have same IP-related characteristics, owners, creation dates etc? Does that mean Cycorp owns UMBEL? Dan, No, it implies that in the UMBEL data space you have equivalence between Classes used to define UMBEL subject concepts (subject matter entities) and OpenCyc. Simple examples via SPARQL against DBpedia using the SPARQL endpoint: http://dbpedia.org/sparql : -- Get DBpedia Entities of Type opencyc:Motorcycle where the dbpedia:name contains pattern: BMW define input:inference 'http://dbpedia.org/resource/inference/rules/umbel#' prefix umbel: http://umbel.org/umbel/sc/ prefix dbpedia: http://dbpedia.org/property/ prefix opencyc: http://sw.opencyc.org/2008/06/10/concept/en/ select ?s where { ?s a opencyc:Motorcycle. ?s dbpedia:name ?n. ?n bif:contains BMW. } -- Get DBpedia Entities of Type umbel:Motorcycle where the dbpedia:name contains pattern: BMW define input:inference 'http://dbpedia.org/resource/inference/rules/umbel#' prefix umbel: http://umbel.org/umbel/sc/ prefix dbpedia: http://dbpedia.org/property/ prefix opencyc: http://sw.opencyc.org/2008/06/10/concept/en/ select ?s where { ?s a umbel:Motorcycle. ?s dbpedia:name ?n. ?n bif:contains BMW. } Both queries return the same results even though in DBpedia you only have entities of type umbel:Motorcycle as shown here: http://dbpedia.org/resource/BMW_R75 . Courtesy of UMBEL and UMBEL inference rules applied to DBpedia, we can now expand the scope of existing DBpedia entities by exploiting relationships that exist across the Classes exposed via rdf:type. Thus, via rdf:type we are able to cross over from the instance realm to the data dictionary realm, and once in the data dictionary realm expand our horizons via class level relationships covering equivalence and the broader and narrower transitive relationships in the Class Hierarchy. So OpenCyc doesn't own UMBEL, it is simply associated with UMBEL by being part of the data dictionary oriented Linked Data Space that UMBEL delivers :-) Kingsley Then between OpenCyc and Wordnet: 1. owl:sameAs Ditto... cheers, Dan -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Role of RDF on the Web and within enterprise applications. was: AW: Berlin SPARQL Benchmark V2 - Results for Sesame, Virtuoso, Jena TDB, D2R Server, and MySQL
the views, thus a high penalty for wrong guess. Andif it is hard enough to figure out where a query plan goes wrong with a given schema, it is harder still to figure it out with a schema that morphs by itself. In the RDB world, for example Oracle recommends saving optimizer statistics from the test environment and using these in the production environment just so the optimizer does not get creative. Now this is the essence of wisdom for OLTP but we are not talking OLTP with RDF. If there is a history of usage and this history is steady and the dba can confirm it as being a representative sample, then automatic materializing of joins is a real possibility. Doing this spontaneously would lead to erratic response times, though. For anything online, the accent is more on predictable throughput than peak throughput. The BSBM query mix does lend itself quite well to automatic materialization but with this, one would not normally do the representation as RDF to begin with, the workload being so typically relational. Publishing any ecommerce database as RDF is of course good but then mapping is the simpler and more predictable route. It is my feeling that RDF has a dual role: 1. interchange format: This is like what XML does, except that RDF has more semantics and expressivity. 2: Database storage format for cases where data must be integrated and is too heterogenous to easily fall into one relational schema. This is for example the case in the open web conversation and social space. The first case is for mapping, the second for warehousing. Aside this, there is potential for more expressive queries through the query language dealing with inferencing, like subclass/subproperty/transitive etc. These do not go very well with SQL views. Orri -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Linked Data Oriented Submissions to the Billion Triples Challenge
All, So far, from the LOD camp, we have: 1. http://www.openlinksw.com/dataspace/oerling/weblog/Orri%20Erling%27s%20Blog/1445 2. http://www.freebase.com/view/user/bio2rdf/public/sparql Stats re. Bio2Rdf.org submission: Here are the final numbers http://www.freebase.com/view/user/bio2rdf/public/sparql 2,345,752,436 or 2.3 Billions triples available from those 27 sparql endpoints http://biocyc.bio2rdf.org/sparql http://biopax.bio2rdf.org/sparql http://chebi.bio2rdf.org/sparql http://cpath.bio2rdf.org/sparql http://dbpedia.bio2rdf.org/sparql http://ec.bio2rdf.org/sparql http://geneid.bio2rdf.org/sparql http://go.bio2rdf.org/sparql http://hgnc.bio2rdf.org/sparql http://homologene.bio2rdf.org/sparql http://inoh.bio2rdf.org/sparql http://iproclass.bio2rdf.org/sparql http://kegg.bio2rdf.org/sparql http://mesh.bio2rdf.org/sparql http://mgi.bio2rdf.org/sparql http://obo.bio2rdf.org/sparql http://omim.bio2rdf.org/sparql http://pdb.bio2rdf.org/sparql http://protein.bio2rdf.org/sparql http://pubchem.bio2rdf.org/sparql http://pubmed.bio2rdf.org/sparql http://reactome.bio2rdf.org/sparql http://taxonomy.bio2rdf.org/sparql http://uniparc.bio2rdf.org/sparql http://uniprot.bio2rdf.org/sparql http://uniref.bio2rdf.org/sparql http://unists.bio2rdf.org/sparql it is 38 public datasources that are globally made available in RDF via Virtuoso SPARQ endpoint. uniref, uniparc and pubmed will be uploaded during the next weeks -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Watchdog.net is plugged into the DBpedia-LOD Cloud
Anja, http://watchdog.net should be added to the new DBpedia-Cloud graphic. It gets to DBpedia via GovTrack linkage. See: http://linkeddata.uriburner.com/about/html/http://watchdog.net/p/nancy_pelosi Note: The value of predicate : http://watchdog.net/about/api#govtrackid is now a GovTrack URI as opposed to the initial literal value. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Watchdog.net is plugged into the DBpedia-LOD Cloud
Dan Brickley wrote: Aaron Swartz wrote: http://www.rdfabout.com/rdf/usgov/congress/people/P000197 which is in my Those are bioguideids, right? Kingsley seems to have some sort of rewrite that converts from govtrack.us URLs to the right thing and since he's the only user, I decided to do it that way. I've lost track of who is behind what here, but looking at above URL I see xmlns:ns=http://xmlns.com/foaf/0.1/; ... Besides FOAF Person, gender, img, which are all fine, I see a foaf:religion unknown or proposed property: ns:religionRoman Catholic/ns:religion ...this isn't in the spec, but I'd consider it for sure if it's proposed on foaf-dev. Also an rdf syntax issue: ns:homepagehttp://www.house.gov/pelosi/ns:homepage should be... ns:homepage rdf:resource=http://www.house.gov/pelosi/ cheers, Dan Dan, Aaron is behind Watchdog.net . Joshua is behind GovTrack. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Watchdog.net is plugged into the DBpedia-LOD Cloud
Joshua Tauberer wrote: Kingsley Idehen wrote: http://watchdog.net should be added to the new DBpedia-Cloud graphic. It gets to DBpedia via GovTrack linkage. See: http://linkeddata.uriburner.com/about/html/http://watchdog.net/p/nancy_pelosi Note: The value of predicate : http://watchdog.net/about/api#govtrackid is now a GovTrack URI as opposed to the initial literal value. Actually the URI isn't quite right. (I can hear Aaron and Kingsley going d'oh after all of the (off-list) emails about this.) Aaron- I suggest that #govtrackid go back to its original value (e.g. 400314) and add a owl:sameAs to http://www.rdfabout.com/rdf/usgov/congress/people/P000197 which is in my URI space for politicians, as opposed to http://www.govtrack.us/congress/person.xpd?id=400314 which is simply the HTML page on GovTrack. Joshua, Why not have both links (one for the page association and owl:sameAs for the Person Entity links) ? The document link is but one association across the data spaces that you an Aaron maintain :-) maybe #govtrackid could become #govtrack_page leaving the URI value unchanged. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: web to semantic web : an automated approach
रविंदर ठाकुर (ravinder thakur) wrote: Hello Friends, I have been following semantic web for some time now and have seen quite a lot of projects being run (dbpedia, FOAF, LOD etc) trying to generate/organize some semantic content. While these approaches might have been successful in their goals, one major problem plaguing semantic web as a whole is the lack of semantic content. Unfortunately there is nothing in sight that we can rely on to generate semantic content for the truckloads of information being put on web everyday. I strongly feel that one of the _wrong_ assumption in semantic web community is that content creators will be creating a semantic data. This I think is too much for the asking from even more technically sound part of web community let along whole of the web community. It hasn't happened over last so many years and I don't see it happening in the near future. To really move the semantic web forward is a mechanism to device a mechanism to _automatcially_ convert the information over the web to semantic information. There are many softwares/services that can be used for this purpose. I am currently developing one prototype for this purpose. This prototype uses services from OpenCalais(http://www.opencalais.com/) to convert ordinary text to semantic form. This service is very limited in what entities supports at the moment but its a very good start. I am pretty sure there will be many other good options available that might be unknown to me. The currently very primitive prototype can be seen at http://arcse.appspot.com. This currently implements very few of the ideas I have for this. This is hosted on Google's AppEngine so sometime gives timeout messages internally so please bear with this :). This automatic conversion however is not a simple task and needs work in lot in domains ranging form NLP to artificial intelligence to semantic web to logic etc. So that's why this mail. I will be more than happy if we can join together to form a like minded team that can work on solving this most important problem plaguing semantic web currently. Waiting for your suggestions/criticisms. And Happy Diwali http://en.wikipedia.org/wiki/Diwali too :) Ravinder Thakur PS : I posted similar query here http://www.nabble.com/web-to-semantic-web-%3A-an-automated-approach-td20064203.htmlas well. That generated some good debate. Ravinder, Please read the various posts I've made about RDF Middleware which is all about the recently branded top-down approach to the Linked Data Web. From day one, I've always subscribed to the belief that structured linked data will need to be generated from existing Web content in order for the Linked Data aspect of the Web to blossom. Links: 1. http://www.openlinksw.com/dataspace/[EMAIL PROTECTED]/weblog/[EMAIL PROTECTED]/1454 2. http://www.openlinksw.com/weblog/public/search.vspx?blogid=127q=rdf%20middlewaretype=textoutput=html 3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtSponger -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Freebase Linked Data
All, A few samplings of what the Freebase Linked Data space offers courtesy of our Linked Data Explorer offerings (Firefox extension [1] or Live hosted service [2]) . 1. http://tinyurl.com/5mhnsu -- About Obama 2. http://tinyurl.com/6br773 -- About Hillary Clinton Note: We connect the container (information resource) with the relevant entities via foaf:primarytopic. Also note the use of UMBEL (but much more to come on this later now that ground zero of the Linked Data has more or less crystallized). 1. http://ode.openlinksw.com 2. http://linkeddata.uriburner.com/ode (*what used to be http://demo.openlinksw.com/rdfbrowser2 | rdfbrowser *) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Linked Data Tutorial Material
Ed Summers wrote: On Fri, Nov 14, 2008 at 6:40 AM, Yves Raimond [EMAIL PROTECTED] wrote: We plan to publish, with Keith, a small how-to for doing a hands-on tutorial like the Web-of-data 101 session we did at the WOD-PD event. With Richard's permission, we'll also take a few things from his session, where he used some of the data we created during our session. The goal of this how-to is to make it easy for people to reproduce the tutorial in different places. This would be extremely helpful to me and some others who are thinking of doing a linked-data tutorial at code4lib2009 [1]. Please post a message here when any of the materials materialize! //Ed [1] http://code4lib.org/2009 All, Shouldn't we be able to use a Wiki to produce and maintain a series of Linked Data Tutorials aimed at a plethora of target audiences? -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
-sc:Graduate * umbel-sc:Communicator * umbel-sc:Person * umbel-sc:Entertainer Links: 1. http://sw.opencyc.com 2. http://umbel.org -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Dbpedia-announcements] ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
Chris Bizer wrote: Hi Andreas, we for sure want to do this, but also did not want to postpone the DBpedia 3.2 release any further. So be ensured that the upcoming public user interface for defining the infobox-to-ontology mappings will include the possibility to reuse existing classes and properties and that external classes and properties will be used within the 3.3 release. Defining the infobox-to-ontology mappings that we currently have was already a lot of work (Anja thanks again), so please be patient with the mappings/reuse of external ontologies. Cheers Chris Andreas, Apropos Chris' comments above, it's coming. DBpedia ontology to UMBEL ontology mapping is one of a number of mappings that will emerge in due course. That said -- by the very nature of this project and RDF based Linked Data in general -- anyone with a vested interest in the ontology mapping efforts can step in and accelerate matters. Thus, I encourage others to participate bearing in mind DBpedia 3.2's increased compatibility with such endeavors. Kingsley -Ursprüngliche Nachricht- Von: Andreas Harth [mailto:[EMAIL PROTECTED] Gesendet: Montag, 17. November 2008 16:55 An: Chris Bizer Cc: public-lod@w3.org; 'Semantic Web'; dbpedia- [EMAIL PROTECTED]; dbpedia- [EMAIL PROTECTED] Betreff: Re: ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase Hi Chris, Chris Bizer wrote: 1. DBpedia Ontology DBpedia now features a shallow, cross-domain ontology, which has been manually created based on the most commonly used infoboxes within Wikipedia great work! One thing: what's the reason for creating your own classes rather than re-using or sub-classing existing ones (foaf:Person, geonames:Feature...)? Same for properties (foaf:name, dc:date...). Regards, Andreas. -- http://swse.deri.org/ - This SF.Net email is sponsored by the Moblin Your Move Developer's challenge Build the coolest Linux based applications with Moblin SDK win great prizes Grand prize is a trip for two to an Open Source event anywhere in the world http://moblin-contest.org/redirect.php?banner_id=100url=/ ___ Dbpedia-announcements mailing list [EMAIL PROTECTED] https://lists.sourceforge.net/lists/listinfo/dbpedia-announcements -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
John Goodwin wrote: Have fun with the new DBpedia knowledge base! Cheers Chris Thanks Chris and team for all your hard work getting this done. I do, however, have a few comments regarding the OWL ontology. I think in general the use of domain and range is perhaps a bit dubious in that for many things I think it is overly specified. I can imagine anyone re-using the Dbpedia properties getting some unexpected inferences from the domain and range restrictions. Also the range restriction seem to be done as an OWL intersection so if, for example, something has a publisher x then x will be inferred to be both a Company and a Person which is probably not what you want. Personally, in all but a few cases, I'd be tempted to generalise or just remove the domain/range restrictions. Any thoughts? john . John, Yes, this issues exist with the DBpedia Ontology, but also bear in mind this is also why UMBEL and DBpedia have been Linked :-) Eventually (as per my other mail posts), the DBpedia Ontology will need to be linked with UMBEL (*which isn't the case right now*) so that it can benefit from the work done in the UMBEL project. That said, UMBEL is also a loosely bound ontology for DBpedia already, so you can also optionally leverage for inference purposes right now. All in all, I am hoping a lot of the inherent flexibilities that the Internet hosted Distributed Database we know as the Web accords start to come through with much more clarity following DBpedia 3.2. Kingsley This email is only intended for the person to whom it is addressed and may contain confidential information. If you have received this email in error, please notify the sender and delete this email which must not be copied, distributed or disclosed to any other person. Unless stated otherwise, the contents of this email are personal to the writer and do not represent the official view of Ordnance Survey. Nor can any contract be formed on Ordnance Survey's behalf via email. We reserve the right to monitor emails and attachments without prior notice. Thank you for your cooperation. Ordnance Survey Romsey Road Southampton SO16 4GU Tel: 08456 050505 http://www.ordnancesurvey.co.uk -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Dbpedia-discussion] ANN: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
the information in any way nor act in reliance on it and notify the sender immediately. Please note that the BBC monitors e-mails sent or received. Further communication will signify your consent to this. - This SF.Net email is sponsored by the Moblin Your Move Developer's challenge Build the coolest Linux based applications with Moblin SDK win great prizes Grand prize is a trip for two to an Open Source event anywhere in the world http://moblin-contest.org/redirect.php?banner_id=100url=/ ___ Dbpedia-discussion mailing list [EMAIL PROTECTED] https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
Jens Lehmann wrote: Hello John, John Goodwin wrote: Thanks Chris and team for all your hard work getting this done. I do, however, have a few comments regarding the OWL ontology. I think in general the use of domain and range is perhaps a bit dubious in that for many things I think it is overly specified. I can imagine anyone re-using the Dbpedia properties getting some unexpected inferences from the domain and range restrictions. Also the range restriction seem to be done as an OWL intersection so if, for example, something has a publisher x then x will be inferred to be both a Company and a Person which is probably not what you want. Personally, in all but a few cases, I'd be tempted to generalise or just remove the domain/range restrictions. Any thoughts? We specified the domains and ranges as disjunctions of classes (not intersection). See the W3C specification of owl:unionOf [1]. The domain and range axioms help to structure DBpedia and clarify the meaning of certain properties. While there is room for improvement, it is not an option to remove all of them. Currently, there are two versions of the infobox extraction: a loose one and a strict one. In the strict one, it is guaranteed that the data complies to the ranges specified in the ontology schema. Currently, only the loose (probably inconsistent) one is provided. Kind regards, Jens [1] http://www.w3.org/TR/owl-guide/#owl_unionOf Jens, What's the URL of the strict one? We are building a DBpedia installer for Virtuoso, so at the very least I want the users of this installer to have choice of strict or loose infobox extraction. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
Georgi Kobilarov wrote: Kingsley, What's the URL of the strict one? We are building a DBpedia installer for Virtuoso, so at the very least I want the users of this installer to have choice of strict or loose infobox extraction. Not publicly available yet. There was a buggy first version of strict, but we decided to no further work on it for release 3.2. Okay. Kingsley Georgi -- Georgi Kobilarov Freie Universität Berlin www.georgikobilarov.com -Original Message- From: Kingsley Idehen [mailto:[EMAIL PROTECTED] Sent: Monday, November 17, 2008 6:14 PM To: Jens Lehmann Cc: public-lod@w3.org; Semantic Web; dbpedia- [EMAIL PROTECTED]; John Goodwin Subject: Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase Jens Lehmann wrote: Hello John, John Goodwin wrote: Thanks Chris and team for all your hard work getting this done. I do, however, have a few comments regarding the OWL ontology. I think in general the use of domain and range is perhaps a bit dubious in that for many things I think it is overly specified. I can imagine anyone re-using the Dbpedia properties getting some unexpected inferences from the domain and range restrictions. Also the range restriction seem to be done as an OWL intersection so if, for example, something has a publisher x then x will be inferred to be both a Company and a Person which is probably not what you want. Personally, in all but a few cases, I'd be tempted to generalise or just remove the domain/range restrictions. Any thoughts? We specified the domains and ranges as disjunctions of classes (not intersection). See the W3C specification of owl:unionOf [1]. The domain and range axioms help to structure DBpedia and clarify the meaning of certain properties. While there is room for improvement, it is not an option to remove all of them. Currently, there are two versions of the infobox extraction: a loose one and a strict one. In the strict one, it is guaranteed that the data complies to the ranges specified in the ontology schema. Currently, only the loose (probably inconsistent) one is provided. Kind regards, Jens [1] http://www.w3.org/TR/owl-guide/#owl_unionOf Jens, What's the URL of the strict one? We are building a DBpedia installer for Virtuoso, so at the very least I want the users of this installer to have choice of strict or loose infobox extraction. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com --- -- This SF.Net email is sponsored by the Moblin Your Move Developer's challenge Build the coolest Linux based applications with Moblin SDK win great prizes Grand prize is a trip for two to an Open Source event anywhere in the world http://moblin-contest.org/redirect.php?banner_id=100url=/ ___ Dbpedia-discussion mailing list [EMAIL PROTECTED] https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Dbpedia-discussion] DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
about a thing from DBpedia and Freebase. For more information about the Freebase links see: http://blog.dbpedia.org/2008/11/15/dbpedia-is-now-interlinked-with- freebase- links-to-opencyc-updated/ 3. Cleaner Abstacts Within the old DBpedia dataset it occurred that the abstracts for different languages contained Wikpedia markup and other strange characters. For the 3.2 release, we have improved DBpedia's abstract extraction code which results in much cleaner abstracts that can safely be displayed in user interfaces. The new DBpedia release can be downloaded from: http://wiki.dbpedia.org/Downloads32 and is also available via the DBpedia SPARQL endpoint at http://dbpedia.org/sparql and via DBpedia's Linked Data interface. Example URIs: http://dbpedia.org/resource/Berlin http://dbpedia.org/page/Oliver_Stone More information about DBpedia in general is found at: http://wiki.dbpedia.org/About Lots of thanks to everybody who contributed to the Dbpedia 3.2 release! Especially: 1. Georgi Kobilarov (Freie Universität Berlin) who designed and implemented the new infobox extraction framework. 2. Anja Jentsch (Freie Universität Berlin) who contributed to implementing the new extraction framework and wrote the infobox to ontology class mappings. 3. Paul Kreis (Freie Universität Berlin) who improved the datatype extraction code. 4. Andreas Schultz (Freie Universität Berlin) for generating the Freebase to DBpedia RDF links. 5. Everybody at OpenLink Software for hosting DBpedia on a Virtuoso server and for providing the statistics about the new Dbpedia knowledge base. Have fun with the new DBpedia knowledge base! Cheers Chris -- Prof. Dr. Christian Bizer Web-based Systems Group Freie Universität Berlin +49 30 838 55509 http://www.bizer.de [EMAIL PROTECTED] - This SF.Net email is sponsored by the Moblin Your Move Developer's challenge Build the coolest Linux based applications with Moblin SDK win great prizes Grand prize is a trip for two to an Open Source event anywhere in the world http://moblin-contest.org/redirect.php?banner_id=100url=/ ___ Dbpedia-discussion mailing list [EMAIL PROTECTED] https://lists.sourceforge.net/lists/listinfo/dbpedia-discussion -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: AW: DBpedia 3.2 release, including DBpedia Ontology and RDF links to Freebase
that are present in a certain dataset. They simply express what's there in the data. In this sense, they are like database schemas: If Publisher has a range of Person, then it means that the publisher *in this particular dataset* is always a person. That's not an assertion about the world, it's an assertion about the dataset. These ontologies are usually not very re-usable. b) Ontologies that are intended as a lingua franca for data exchange between different applications. They are designed for broad re-use, and thus usually do not add many restrictions. In this sense, they are more like controlled vocabularies of terms. Dublin Core is probably the prototypical example, and FOAF is another good one. They usually don't allow as many interesting inferences. I think that these two kinds of ontologies have very different requirements. Ontologies that are designed for one of these roles are quite useless if used for the other job. Ontologies that have not been designed for either of these two roles usually fail at both. Returning to DBpedia, my impression is that the DBpedia ontology is intended mostly for the first role. Maybe it should be understood more as a schema for the DBpedia dataset, and not so much as a re-usable set of terms for use outside of the Wikipedia context. (I might be wrong, I was not involved in its creation.) Richard -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Suggestions/existing material for Linked Data tutorial?
Yves Raimond wrote: Hello! We plan to publish, with Keith, a small how-to for doing a hands-on tutorial like the Web-of-data 101 session we did at the WOD-PD event. With Richard's permission, we'll also take a few things from his session, where he used some of the data we created during our session. The goal of this how-to is to make it easy for people to reproduce the tutorial in different places. I (finally) started this linked data tutorial how-to, at: http://community.linkeddata.org/MediaWiki/index.php?WODPDTutorial Still a lot to do though - I didn't even touch the second part of the tutorial. I also wrote a small script, linked from this Wiki page, and implementing a small linked data server [1] - attendees can POST RDF data to it, and it gets published as linked data. It also provides a SPARQL end-point. Anyway, early comments are welcome :-) I hope to finish that soon (I need to do this tutorial at the BBC soonish, so it'll have to be ready by then :-) ) Cheers! y [1] http://moustaki.org/resources/rdf-pastebin.tar.gz Yves, Nice start. We will add our own tutorial which starts from: Get Yourself a URI in 5 mins of less :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Size matters -- How big is the danged thing
David Wood wrote: Sorry to intervene here, but I think Kingsley's suggestion sets up a false dicotomy. REST principles (surely part of everything we stand for :) suggest that the source of RDF doesn't matter as long as a URL returns what we want. Late binding means not having to say you're sorry. Is it a good idea to set up a class system where those who publish to files are somehow better (or even different!) than those who publish via adapters? David, Yes, the dichotomy is false if the basis is: Linked Data irrespective of means or source, as long as the URIs are de-referencable. On the other hand, if Linked Data generated on the fly isn't deemed part of the LOD cloud (the qualm expressed in Giovanni's comments) then we have to call RDF-ized Linked Data something :-) You can count the warehouse (an arrive at hub size) but the RDF-ized stuff is a complete red herring (imho - see cool fractal animations post). What I am hoping is a more interesting quesion is this: have we reached the point were we can drop burgeoning from the state of the Linked Data Web? Do we have a hub that provides enough critical mass for the real fun to start (i.e., finding stuff with precision that data object properties accord) ? Personally, I think the Linked Data Web has reached this point, so our attention really has to move more towards showing what Linked Data adds to the Web in general. Kingsley So, I vote for counting all of it. Isn't that what Google and Yahoo do when they count the number of pages indexed? Regards, Dave -- On Nov 21, 2008, at 4:26 PM, Kingsley Idehen [EMAIL PROTECTED] wrote: Giovanni Tummarello wrote: Overall, that's about 17 billion. IMO considering myspace 12 billion triples as part of LOD, is quite a stretch (same with other wrappers) unless they are provided by the entity itself (E.g. i WOULD count in livejournal foaf file on the other hand, ok they're not linked but they're not less useful than the myspace wrapper are they? (in fact they are linked quite well if you use the google social API) Giovanni Giovanni, Maybe we should use the following dichotomy re. the Web of Linked Data (aka. Linked Data Web): 1. Static Linked Data or Linked Data Warehouses - which is really what the LOD corpus is about 2. Dynamic Linked Data - which is what RDF-zation middleware (including wrapper/proxy URI generators) is about. Thus, I would say that Jim is currently seeking stats for the Linked Data Warehouse part of the burgeoning Linked Data Web. And hopefully, once we have the stats, we can get on to the more important task of explaining and demonstrating the utility of the humongous Linked Data corpus :-) ESW Wiki should be evolving as I write this mail (i.e. tabulated presentation of the data that's already in place re. this matter). All: Could we please stop .png and .pdf based dispatches of data, it kinda contradicts everything we stand for :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Size matters -- How big is the danged thing
Aldo Bucchi wrote: On Fri, Nov 21, 2008 at 7:51 PM, Kingsley Idehen [EMAIL PROTECTED] wrote: Yves Raimond wrote: On Fri, Nov 21, 2008 at 8:08 PM, Giovanni Tummarello [EMAIL PROTECTED] wrote: Overall, that's about 17 billion. IMO considering myspace 12 billion triples as part of LOD, is quite a stretch (same with other wrappers) unless they are provided by the entity itself (E.g. i WOULD count in livejournal foaf file on the other hand, ok they're not linked but they're not less useful than the myspace wrapper are they? (in fact they are linked quite well if you use the google social API) Actually, I don't think I can agree with that. Whether we want it or not, most of the data we publish (all of it, apart from specific cases e.g. review) is provided by wrappers of some sort, e.g. Virtuoso, D2R, P2R, web services wrapper etc. Hence, it makes not sense trying to distinguish datasets on the basis they're published through a wrapper or not. Within LOD, we only segregate datasets for inclusion in the diagram on the basis they are published according to linked data principles. The stats I sent reflect just that: some stats about the datasets currently in the diagram. The origin of the data shouldn't matter. The fact that it is published according to linked data principles and linked to at least one dataset in the cloud should matter. Giovanni Yves, I agree. But I am sure you can also see the inherent futility in pursuing the size of the pure Linked Data Web :-) The moment you arrive at a number it will be obsolete :-) I would frame the question this way: is LOD hub now dense enough for basic demonstrations of Linked Data Web utility to everyday Web users? For example, can we Find stuff on the Web with levels of precision and serendipity erstwhile unattainable? Can we now tag stuff on the Web in a manner that makes tagging useful? Can we alleviate the daily costs of Spam on mail inboxes? Can all of the aforementioned provide the basis for relevant discourse discovery and participation? Sorry, this is getting too interesting to stay in lurker mode ;) Kingsley, absolutely. We have got to that point. The fun part has begun. To quote Jim, who started this thread: http://blogs.talis.com/nodalities/2008/03/jim_hendler_talks_about_the_se.php Go to minute 28 aprox ( I can't listen to it here, I just blocked mp3's ). Jim touches on how a geo corpus can be used to dissambiguate tags on flickr. This is one such use, low hanging fruit wrt the huge amount of linked data, and a first timer in terms of IT. This was not possible last year! It is now. I guess that is THE question now: What can we do this year that we couldn't do last year? ( thanks to the massive amount of available LOD ). Best, A Aldo, Yep! So we should start building up a simple collection (in a Wiki) of simple and valuable things you can now achieve courtesy of Linked Data :-) Find replacing Search as the apex of the Web value proposition pyramid for everyday Web Users. Courtesy of Linked Data (warehouse and/or dynamic), every Web information resource is now a DBMS View in disguise :-) Kingsley -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Size matters -- How big is the danged thing
David Wood wrote: On Nov 21, 2008, at 5:51 PM, Kingsley Idehen wrote: I would frame the question this way: is LOD hub now dense enough for basic demonstrations of Linked Data Web utility to everyday Web users? For example, can we Find stuff on the Web with levels of precision and serendipity erstwhile unattainable? Can we now tag stuff on the Web in a manner that makes tagging useful? Can we alleviate the daily costs of Spam on mail inboxes? Can all of the aforementioned provide the basis for relevant discourse discovery and participation? An interesting experiment might be to start at some bit of RDF (a FOAF document or some such) and follow-your-nose from link to see to see how far the longest path is. If it is very, very long (maybe even nicely loopy since the LOD effort), then life is good. Regards, Dave Dave, That's what this is all about: http://b3s.openlinksw.com/ ( a huge Linked Data corpus, talking 11 Billion or so triples). What was missing from this demo all along was a Find feature that hides all the SPARQL :-) Also, there will be more when we finally release the long overdue update to the OpenLink Data Explorer :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: linked data mashups
Peter Ansell wrote: 2008/11/23 Juan Sequeda [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] Hi Giovanni and all On Sat, Nov 22, 2008 at 7:33 PM, Giovanni Tummarello [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote: I guess that is THE question now: What can we do this year that we couldn't do last year? ( thanks to the massive amount of available LOD ). Two days ago the discussion touched this interesting point. I do not know how to answer this question. Ideas? We need to start consuming linked data and making reall mashup applications powered by linked data. A couple of days I just mentioned the link for SQUIN: http://squin.sourceforge.net/ The idea of SQUIN came out of ISWC08 with Olaf Hartig. The objective is to make LOD accesible easily to web2.0 app developers. We envision adding an S compoment to the LAMP stack. This will allow people to easily query LOD from their own server. We should have a demo ready in the next couple of weeks. We believe that this is something needed to actually start using LOD and making it accesible to everybody. How does SQIUN differ to a typical HTTP SPARQL endpoint? So far it accepts a query parameter as a SPARQL select statement and executes the parameter on (some configured?) SPARQL endpoints from looking at the single sourcefile I could find [1]. Having said that, I have been holding off getting my bio2rdf server to actually process rdf but it doesn't look so hard now. (The bio2rdf server is actually more generic than just biology or even bio2rdf but it is still named that in response to its origins. And in contrast to SQUIN it focuses on CONSTRUCT queries rather than SELECT) On the subject of mashups I have been thinking in the last few days of combining the bio2rdf server with the pipes.deri.org http://pipes.deri.org interface for mashups, as some fairly sophisticated mashups can be done on pipes.deri.org http://pipes.deri.org, but a lot of the generic queries seem to be better handled at the client level where people can control with configurations what endpoints are used and have backups if a particular endpoint fails. Cheers, Peter [1] http://tinyurl.com/6cvdl8 Peter, Has anything happened re. cross-linking the data across bio2rdf.org and dbpedia.org? Sane cross-linking is vital to Linked Data Web oriented Meshups. Note, there is a distinct difference between a Mashup and a Meshup in my world view. Mashups are nice looking opaque Web pages that have code behind them while Meshups are transparent Web pages with Linked Data behind them (i.e. the data object URIs are accessible to machines). A Meshup style page is really the Linked Data Web's equivalent of a traditional DBMS View. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: linked data mashups
Peter Ansell wrote: 2008/11/24 Kingsley Idehen [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] Peter Ansell wrote: 2008/11/23 Juan Sequeda [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED] Hi Giovanni and all On Sat, Nov 22, 2008 at 7:33 PM, Giovanni Tummarello [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote: I guess that is THE question now: What can we do this year that we couldn't do last year? ( thanks to the massive amount of available LOD ). Two days ago the discussion touched this interesting point. I do not know how to answer this question. Ideas? We need to start consuming linked data and making reall mashup applications powered by linked data. A couple of days I just mentioned the link for SQUIN: http://squin.sourceforge.net/ The idea of SQUIN came out of ISWC08 with Olaf Hartig. The objective is to make LOD accesible easily to web2.0 app developers. We envision adding an S compoment to the LAMP stack. This will allow people to easily query LOD from their own server. We should have a demo ready in the next couple of weeks. We believe that this is something needed to actually start using LOD and making it accesible to everybody. How does SQIUN differ to a typical HTTP SPARQL endpoint? So far it accepts a query parameter as a SPARQL select statement and executes the parameter on (some configured?) SPARQL endpoints from looking at the single sourcefile I could find [1]. Having said that, I have been holding off getting my bio2rdf server to actually process rdf but it doesn't look so hard now. (The bio2rdf server is actually more generic than just biology or even bio2rdf but it is still named that in response to its origins. And in contrast to SQUIN it focuses on CONSTRUCT queries rather than SELECT) On the subject of mashups I have been thinking in the last few days of combining the bio2rdf server with the pipes.deri.org http://pipes.deri.org http://pipes.deri.org interface for mashups, as some fairly sophisticated mashups can be done on pipes.deri.org http://pipes.deri.org http://pipes.deri.org, but a lot of the generic queries seem to be better handled at the client level where people can control with configurations what endpoints are used and have backups if a particular endpoint fails. Cheers, Peter [1] http://tinyurl.com/6cvdl8 Peter, Has anything happened re. cross-linking the data across bio2rdf.org http://bio2rdf.org and dbpedia.org http://dbpedia.org? I have been waiting for information about what progress has been made with the community based infobox extraction framework. Then the relevant predicates in the protein/gene/chemical infoboxes can be used pretty easily for linkages. So I am assuming this hasn't happened, based on your response? Sane cross-linking is vital to Linked Data Web oriented Meshups. Note, there is a distinct difference between a Mashup and a Meshup in my world view. Mashups are nice looking opaque Web pages that have code behind them while Meshups are transparent Web pages with Linked Data behind them (i.e. the data object URIs are accessible to machines). A Meshup style page is really the Linked Data Web's equivalent of a traditional DBMS View. I do understand the difference, but I tend to use the term mashup for any combining of the data sources independent of the presentation. Its hard enough defining a mashup when people ask for a definition without going for another similar term from my experience. Labels are very secondary in how I tend to look at things. In due course the difference between Mashing and Meshing will be self evident, especially if we get all the major Linked Data hubs connected properly. Kingsley Cheers, Peter -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: linked data hosted somewhere
रविंदर ठाकुर (ravinder thakur) wrote: hi i am trying to find some working service hosting the current linkedopendata and providing some programming interface. does any such service exists at the moment ? thanks ravinder thakur Ravinder, Could you be a little more specific? :-) For instance, do you mean: 1. A service that provides SPARQL protocol access to all the warehoused data from the current LOD cloud? 2. ditto but using a service that is higher level than SPARQL? -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: linked cloud data hosted somewhere
Ravinder, i am looking for a service which provide LOD with _any_ protocol. even a low level query system like sparql will be awesome!!! Here is what we have in the making: 1. DBpedia SPARQL endpoint on EC2 (which simply means Virtuoso DB in the Clouds with DBpedia loaded, configured, and optimized) - ETA 24 hrs or less (we are just completing a number of reload runs so we have accurate sense of how long it takes to assemble DBpedia from a Virtuoso backup on EC2; current estimate is 1 1.5 hrs) 2. Rest of the LOD cloud warehouse in bits (i.e. one data space at a time like DBpedia with a sparql endpoint for each) 3. A Virtuoso Cluster edition with the entire LOD cloud warehouse pre loaded + sparql endpoint (our Billion Triples Challenge demo [2] is the precursor of this deliverable). We will announce the availability and provide installation guides for each of these linked data spaces in the clouds post AMI construction and commissioning. Requirements: 1. Initialize an instance of the Virtuoso Universal Server paid AMI [2] Links: 1. http://b3s.openlinksw.com/ 2. https://aws-portal.amazon.com/gp/aws/user/subscription/index.html?offeringCode=6CB89F71 -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: linked data hosted somewhere
Hugh Glaser wrote: Thanks, Kingsley and Aldo. I have to say you raise quite a lot of concerns, or at least matters of interest. I really don't think it is a big deal that I asked someone to consider resources when accessing my web site, and I am a bit uncomfortable that I then get messages effectively telling me that my software is poor and I should be using (buying?) something else. Hugh, You're losing me a little, I don't think Aldo or I were making any comments about your software per se. or making suggestions about alternatives. Anyway more comments inline below. On 26/11/2008 02:12, Kingsley Idehen [EMAIL PROTECTED] wrote: Hugh Glaser wrote: I thought that might be the answer. So what is the ontology of the error, so that my SW application can deal with it appropriately? If it ain¹t RDF it ain¹t sensible in the Semantic Web. ;-| And the ³entitlement² to spend lots of money by accident; a bit worrying, although I assume there are services that allow me to find out at least estimates of the cost. If you are querying via iSQL or the Virtuoso Conductor you wont be moving lots of data between your desktop and EC2. If you do large constructs over the sparql protocol or anything else that produces large HTTP workloads between EC2 and your location, then you will incur the costs (btw - Amazon are quite aggressive re. the costs, so you really have to be serving many client i.e., offering a service for costs being a major concern). Er, yes, that was the question we were discussing. Large constructs over the sparql prototcol. With respect to costs, I never mentioned Amazon, so I am not sure why that is the benchmark for comparison. But I don't want to have a go at the Openlink software (I often recommend it to people); I was just asking about limitations. All software has limitations. Anyway, Virtuoso let's you control lots of things, including shutting down the sparql endpoint. In addition, you will soon be able to offer OAuth access to sparql endpoint etc.. Yes, and I didn't really want to have the overhead of interacting with Ravinder to explain why I had shut down his access to the SPARQL endpoint. I suspect that your comment about a bill is a bit of a joke, in that normal queries do not require money? But it does raise an interesting LOD question. Ravinder asked for LOD sets; if I have to pay for the query service, is it LOD? You pay for traffic that goes in and out of your data space. (effective November 26, 2008) Fixed Costs ($) snip amazon costs/ Here is a purchase link that also exposes the items above. https://aws-portal.amazon.com/gp/aws/user/subscription/index.html?ie=UTF8offe ringCode=6CB89F71 Of course, you can always use the Open Source Edition as is and reconstruct DBpedia from scratch, the cost-benefit analysis factors come down to: 1. Construction and Commissioning time (1 - 1.5 hrs vs 16 - 22 hrs) 2. On / Off edition variant of live DBpedia instance that's fully tuned and in sync with the master Getting back to dealing with awkward queries. Detecting what are effectively DoS attacks is not always the easiest thing to do. Has Bezzo really solved it for a SPARQL endpoint while providing a useful service to users with a wide variety of requirements? I believe so based on what we can do with Virtuoso on EC2. One major example is the backup feature where we can sync from a Virtuoso instance into S3 buckets. Then perform a restore from those buckets (what we do re. DBpedia). In our case we offer HTTP/WebDAV or the S3 protocol for bucket access. I don't think this contributes to helping to service complex SPARQL queries, or have I missed somthing? Hugh: I certainly had my response above a little tangled :-( To clarify, re. Bezos and DOS bit. 1. EC2 instances can be instantiated and destroyed at will 2. Virtuoso (and I assume other SPARQL engines) have DOS busting features such as SPARQL Query Cost Analysis and Rate Limits for HTTP requests. In fact, people don¹t usually offer open SQL access to Open Databases for exactly this reason. I like to think the day will come when the Semantic Web is so widely used that we will have the same problem with SPARQL endpoints. The Linked Data Web is going to take us way beyond anything SQL could even fantasize about (imho). And one such fantasy is going to be accessible sparql endpoints without bringing the house down :-) Now there I agree. The power of LD/SW or whatever you call it will indeed take us a long way further. And I agree on the fantasy, which is actually what I was saying all along. It is a fantasy to suggest that you can do all the wrong you want. Exactly! But I think it is sensible to take the question to a new thread... No problem :-) Kingsley Best Hugh Kingsley -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web
Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)
reply. My point was simply that, if you want to push things over the edge, why not get your own box. We all take care of our infrastructure and know its limitations. So, I formally apologize. I am by no means endorsing one piece of software over another ( save for mine, but it does't exist yet ;). My preferences for virtuoso come from experiential bias. I hope this clears things up. I apologize for the traffic. However, I do make a formal request for some sense of humor. This list tends to get into this kind of discussions, and we will start getting more and more visits from outsiders who are not used to this sort of sharpness. Best, A -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)
://quebec.bio2rdf.org/download/virtuoso/indexed/ [3] http://quebec.bio2rdf.org/download/n3/ [4] http://sourceforge.net/project/platformdownload.php?group_id=142631 [5] http://bio2rdf.mquter.qut.edu.au/ --~--~-~--~~~---~--~~ You received this message because you are subscribed to the Google Groups bio2rdf group. To post to this group, send email to [EMAIL PROTECTED] To unsubscribe from this group, send email to [EMAIL PROTECTED] For more options, visit this group at http://groups.google.com/group/bio2rdf?hl=en -~--~~~~--~~--~--~--- -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Can we afford to offer SPARQL endpoints when we are successful? (Was linked data hosted somewhere)
Danny Ayers wrote: %=profanity /% or something - cool thread while I disagree with many of Aldo's individual points, getting them surfaced is really positive in response to a line from the firestarter: The only reason anyone can afford to offer a SPARQL endpoint is because it doesn't get used too much? while my love of SPARQL is enormous, I can't see the SPARQL endpoint being a lasting scenario. But you seem to assume that SPARQL endpoints are the definitive scenario today? I see SPARQL endpoints as simply optional. The public visibility will be determined by needs and the wishes of a give Linked Data Space owner. Take DBpedia as an example, it works fine as a Linked Data Space on its own, but it also offers a SPARQL endpoint for those who want to speak SPARQL. As I've stated (already), the real issue comes down to deep DBMS level issues re. scalability and query processing prowess. Today, we don't have a ODBC or JDBC DSNs published on the Web primarily because the naming granularity isn't there (i.e. record level naming), single record access and manipulation has always been a problem for SQL, and there's never been a truly standardized data access mechanism that incorporates the networking layer. Linked Data picks up from where the RDBMS realm simply runs out of steam re. Open Data Access, but this doesn't mean that the techniques in the RDBMS realm can form the basis of a new kind of DBMS frontier of the kind I believe the Linked Data Web ultimately delivers. linking and the fresh approach to caching this will demand, need another rev. before the web starts doing data efficiently When you say: ..Web doing data efficiently, what do you mean? Scalable and change sensitive data access on the Linked Data Web is fundamentally an issue for Linked Data Deployment platform player differentiation (imho). I think HTTP, RDF, and SPARQL (plus extensions) already provide ample building blocks for building solid platforms. We just have to remember that data access and DBMS technology predates the Web and lots of the techniques and knowledge from this realm are vital when it comes to scalability and efficiency challenges. In my eyes, and from my experience, the Web is only beginning to comprehend DBMS matters. And like most things relating to humans, some lessons are only learned the hard way :-) Kingsley the answer to the quoted line is the question - how can you not afford? Classic stuff re. amazon opening up their silo a little bit - guess what, profit! pip, Danny. 2008/11/28 Juan Sequeda [EMAIL PROTECTED]: On Thu, Nov 27, 2008 at 2:33 PM, Peter Ansell [EMAIL PROTECTED] wrote: 2008/11/27 Richard Cyganiak [EMAIL PROTECTED] Hugh, Here's what I think we will see in the area of RDF publishing in a few years: - those query capabilities are described in RDF and hence can be invoked by tools such as SQUIN/SemWebClient to answer certain queries efficiently I still don't understand what SQUIN etc have that goes above Jena/Sesame/SemWeb etc which can do this URI resolution with very little programming knowledge in custom applications. True, jena/sesame does everything that SQUIN entails to do. However, SQUIN is oriented to the web2.0 developers. How is a php/ror web developer going to interact with the web of data and make some kind of semantic-linked data mashup over a night? SQUIN will let them do this. No need of having jena, learning jena, etc. Make it simple! If it is not simple, then developers are not going to use it. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: DBpedia+Virtuoso EC2 AMI
Dan Brickley wrote: Kingsley Idehen wrote: All, Quick FYI. What is Virtuoso+DBpedia AMI for EC2? An pre-installed and fully tuned edition of Virtuoso that includes a fully configured DBpedia instance on Amazon's EC2 Cloud platform. Nice work :) Benefits? Generally, it provides a no hassles mechanism for instantiating personal, organization, or service specific instances of DBpedia within approximately 1.5 hours as opposed to a lengthy rebuild from RDF source data that takes between 8 - 22 hours depending on machine hardware configuration and host operating system resources. From a Web Entrepreneur perspective it offers all of the generic benefits of a Virtuoso AMI on EC2 plus the following: 1. Instant bootstrap of a dense Lookup Hub for Linked Data Web oriented solutions 2. No exposure to any of the complexities and nuances associated with deployment of de-referencable URIs (you have a local DBpedia replica) Does this mean all the URIs are dbpedia.mysite.example.com rather than dbpedia.org/* ? In http://kingsley.idehen.name/page/Linked_Data I see 'http://dbpedia.org/resource/Linked_Data' at the top of the page, but no owl:sameAs to http://kingsley.idehen.name/page/Linked_Data Template bug. It should read: About: http://kingsley.idehen.name/resource/Linked_Data Then hook back to DBpedia via owl:sameAs and/or an some atrribution tiples. The raw data links aren't working for me at the moment: eg. curl http://kingsley.idehen.name/sparql?default-graph-uri=http%3A%2F%2Fdbpedia.orgquery=DESCRIBE+%3Chttp://kingsley.idehen.name/resource/Linked_Data%3Eoutput=xml; ?xml version=1.0 encoding=utf-8 ? rdf:RDF xmlns:rdf=http://www.w3.org/1999/02/22-rdf-syntax-ns#; xmlns:rdfs=http://www.w3.org/2000/01/rdf-schema#; Template bug. etc., ... but I get the impression the RDF files use the main dbpedia URI space, while the HTML uses local references. If that's the case, how might this look in RDFa? It should be de-referencing the http://kingsley.idehen.name/resource/Linked_Data URI locally since the DBpedia data is in a local graph with IRI http://dbpedia.org . The routes back to the main/public DBpedia corpus / space is to be via owl:sameAs or Attribution triples. Re. RDFa we will can put this in the template once this first cut settles. I want to get DBpedia replicated with ease for personal and service specific use as task 1 :-). Kingsley cheers, Dan 3. Predictable performance and scalability due localization of query processing (you aren't sharing the public DBpedia server with the rest of the world) Features: 1. DBpedia public instance functionality replica (re. RDF and (X)HTML resource description representations SPARQL endpoint) 2. Local URI de-referencing (so no contention with public endpoint) and part of the Linked Data Deployment 3. Fully tuned Virtuoso instance for DBpedia data set hosting 1. http://kingsley.idehen.name/resource/Linked_Data 2. http://kingsley.idehen.name/resource/Entity-attribute-value_model 3. http://kingsley.idehen.name/resource/Hyperdata 4. http://kingsley.idehen.name/resource/Object_hyperlinking 5. http://kingsley.idehen.name/resource/Barack_Obama -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Potential Home for LOD Data Sets
Hugh Glaser wrote: Exciting stuff, Kingsley. I'm not quite sure I have worked out how I might use it though. The page says that hosting data is clearly free, but I can't see how to get at it without paying for it as an EC2 customer. Is this right? Cheers Hugh, No, shouldn't cost anything if the LOD data sets are hosted in this particular location :-) Kingsley Hugh On 01/12/2008 15:30, Kingsley Idehen [EMAIL PROTECTED] wrote: All, Please see: http://aws.amazon.com/publicdatasets/ ; potentially the final destination of all published RDF archives from the LOD cloud. I've already made a request on behalf of LOD, but additional requests from the community will accelerate the general comprehension and awareness at Amazon. Once the data sets are available from Amazon, database constructions costs will be significantly alleviated. We have DBpedia reconstruction down to 1.5 hrs (or less) based on Virtuoso's in-built integration with Amazon S3 for backup and restoration etc.. We could get the reconstruction of the entire LOD cloud down to some interesting numbers once all the data is situated in an Amazon data center. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Potential Home for LOD Data Sets
Hugh Glaser wrote: Thanks for the swift response! I'm still puzzled - sorry to be slow. http://aws.amazon.com/publicdatasets/#2 Says: Amazon EC2 customers can access this data by creating their own personal Amazon EBS volumes, using the public data set snapshots as a starting point. They can then access, modify and perform computation on these volumes directly using their Amazon EC2 instances and just pay for the compute and storage resources that they use. Does this not mean it costs me money on my EC2 account? Or is there some other way of accessing the data? Or am I looking at the wrong bit? Okay, I see what I overlooked: the cost of paying for an AMI that mounts these EBS volumes, even though Amazon is charging $0.00 for uploading these huge amounts of data where it would usually charge. So to conclude, using the loaded data sets isn't free, but I think we have to be somewhat appreciative of a value here, right? Amazon is providing a service that is ultimately pegged to usage (utility model), and the usage comes down to value associated with that scarce resource called time. Ie Can you give me a clue how to get at the data without using my credit card please? :-) You can't you will need someone to build an EC2 service for you and eat the costs on your behalf. Of course such a service isn't impossible in a Numerati [1] economy, but we aren't quite there yet, need the Linked Data Web in place first :-) Links: 1. http://tinyurl.com/64gsan Kingsley Best Hugh On 05/12/2008 02:28, Kingsley Idehen [EMAIL PROTECTED] wrote: Hugh Glaser wrote: Exciting stuff, Kingsley. I'm not quite sure I have worked out how I might use it though. The page says that hosting data is clearly free, but I can't see how to get at it without paying for it as an EC2 customer. Is this right? Cheers Hugh, No, shouldn't cost anything if the LOD data sets are hosted in this particular location :-) Kingsley Hugh On 01/12/2008 15:30, Kingsley Idehen [EMAIL PROTECTED] wrote: All, Please see: http://aws.amazon.com/publicdatasets/ ; potentially the final destination of all published RDF archives from the LOD cloud. I've already made a request on behalf of LOD, but additional requests from the community will accelerate the general comprehension and awareness at Amazon. Once the data sets are available from Amazon, database constructions costs will be significantly alleviated. We have DBpedia reconstruction down to 1.5 hrs (or less) based on Virtuoso's in-built integration with Amazon S3 for backup and restoration etc.. We could get the reconstruction of the entire LOD cloud down to some interesting numbers once all the data is situated in an Amazon data center. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Exercise: LOD questions (R)†was ( Do we need another list(s)? )
do it :). I dont think its a coincidence that some of the smartest people who worked on semantic web now work for them. (but of course there is much more than a good idea for a successful business so they might go bust anyway obviously) A final quote people like me don't a) know about this and b) don't understand how to use it once they do? I would say some additional education is necessary to make this understood... i would also say that in a broader sense the semantic web message has gotten lost under a mass of acryonyms and theory for a more articulate attept at an explanation of what happened i agree a lot with this post http://inamidst.com/whits/2008/technobunkum by Sean Palmer I dont think more education is needed Juan, one really should teach something if .. the answer is known else its called brainstorming or handwaving (according to weather you're in good faith or not) note that this is all but a bashing on the power of handling loosely structured data and RDF. I think on the other hand RDFa will triumph and so people will be probably making their own little vucabolaries.. but starting from the web 2.0 approach and practical how do i bring visitors, how to do simple site to site integration use cases. Giovanni -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Potential Home for LOD Data Sets - is it Open?
Hugh Glaser wrote: Thanks Kingsley. Getting there, I think/hope. So exactly what is the URI? I run something like select * where { ?s ?p ZNF492 } and get back things like http://purl.org/commons/record/ncbi_gene/57615, but these are not URIs in the Amazon cloud, and so if that is where I was serving my Linked Data from, they are not right. Hugh, You are experimenting with a work in progress. DBpedia on EC2 is the showcase for now. http://kingsley.idehen.name/resource/Linked_Data will de-reference locally and produce triples that connect to http://dbpedia.org/resource/Linked_Data. But note, http://kingsley.idehen.name is currently serving up neurocommons data while finish what I am doing with neurocommons. Would it look something like http://ec2-67-202-37-125.compute-1.amazonaws.com/record/ncbi_gene/57615 or something else? Yes, and like the DBpedia example link back to the public neurocommons URIs. Or is it just that neurocommons is not offering resolvable URIs on the EC2 (if I understand the term), but they could switch on something (in Virtuoso?) that would give me back resolvable URIs on Amazon? The instance on EC2 will do what I stated above once we are done with the de-referencing rules construction and verification etc.. And I am also now wondering who pays when I curl the Amazon URI? It can't be me, as I have no account. The person or organization deploying the linked data pays the bill for the computing resources by amazon and the linked management and deployment services from Virtuoso. It isn't the person who put the data there, as you said it was being hosted for free. Be careful here, the hosting issue was simply about an additional home for the RDF data set archives. The place from with Quad / Triple stores load their data en route to Linked Data publishing (via EC2 or somewhere else). In the case of an EC2 AMI the cost of loading from an S3 Bucket into an EC2 AMI is minimal (if anything at all) since the data is in the same data center as the EC2 AMI. I assume that it means that it must be the EC2 owner, who is firing up the Virtuoso magic to deliver the RDF for the resolved URI? Yes, and in the case of Virtuoso you simply have a platform that offers Linked Data Deployment with a local de-referencing twist while retaining links to original URIs (as per my DBpedia example). Once I am done with neurocommons I'll temporary put DBpedia and Neurocommons on different ports at http://kingsley.idehen.name for demo purposes :-) Kingsley Best Hugh On 07/12/2008 03:34, Kingsley Idehen [EMAIL PROTECTED] wrote: Hugh Glaser wrote: Thanks Kingsley. In case I am still misunderstanding, a quick question: On 06/12/2008 23:53, Kingsley Idehen [EMAIL PROTECTED] wrote: ... Linking Open Data Sets on the Web, is about publishing RDF archives with the following characteristics: 1. De-referencable URIs ... So if someone decides to follow this way and puts their Linked Data in the Amazon cloud using this method, can I de-reference a URI to it using my normal browser or curl it from my machine? Hugh, Absolutely! For instance, a EC2 based instance of DBpedia will do the following: 1. Localize the de-referencing task (i.e. not pass this on to general public instance of DBpedia) 2. Project triples that connect back to the http://dbpedia.org via owl:sameAs (*this was basically what Dan was clarifying in our exchange earlier this week*) The fundamental goal is to use Federation to propagate Linked Data (meme, value prop., and business models) :-) btw - Neurocommons is a data set is now live at the following locations: 1. http://kingsley.idehen.name (*temporary as I simply used this to set up the AMI and verify the entire DB construction process) 2. http://ec2-67-202-37-125.compute-1.amazonaws.com/ (*instance set up by Hugh to double-verify what I did*) Neurcommons takes about 14hrs+ to construct under the best of circumstances. The process is now 1.15 hrs and you have your own personal or service specific neurocommons database. Next stop, Bio2Rdf :-) Thanks. Hugh -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: The next Internet giant: linking open data, providing open access to repositories
Shavkat Karimov wrote: I think people wouldn't pay any fees to search the web. I don't believe I said or implied they would. The FREEdom of the Internet is why it is so powerful. Nobody is disputing that. Also, while I agree with most other Milton's statements, this new Internet giant might still be Google, since they are already going to this direction in many ways. Thanks for the insightful post! Again, the issue of giant/behemoths is somewhat irrelevant (imho). The issue is where and how the Web evolves. I believe we are adding the property oriented Find of the Linked Data Web aspect to the full text pattern oriented Search of the Document Web aspect. The Web is simply evolving and so will the value it delivers etc.. Federation is making a comeback courtesy of the Linked Data aspect of the Web :-) Kingsley Kindest regards, Shavkat http://www.seomanager.com --- Monday, December 8, 2008, 12:14:05 AM, you wrote: Sw-MetaPortal-ProjectParadigm wrote: The next Internet giant company will be linking open data and providing open access to repositories, in the process seamlessly combining both paid for subscriptions, Creative Commons or similar license based or open source software schemes. Revenues will be generated among other things from online advertising streams currently not utilized by Google or Yahoo! In the big scheme of things this company will redefine the concept of internet search to provide access to deep(er) web levels of data and information for which users will be willing to pay an annual flat fee subscription. Sound improbable? Non-profit organizations dedicated to providing global open access will soon start exploring just such business schemes to determine if it is feasible to fund and maintain the server farms, hard and software to do just that. Milton, Not improbable, that's what coming :-) The only tweak of you statement would be this: there will be a federation of linked data spaces rather than a single behemoth :-) Kingsley Milton Ponson GSM: +297 747 8280 Rainbow Warriors Core Foundation PO Box 1154, Oranjestad Aruba, Dutch Caribbean www.rainbowwarriors.net Project Paradigm: A structured approach to bringing the tools for sustainable development to all stakeholders worldwide www.projectparadigm.info NGO-Opensource: Creating ICT tools for NGOs worldwide for Project Paradigm www.ngo-opensource.org MetaPortal: providing online access to web sites and repositories of data and information for sustainable development www.metaportal.info SemanticWebSoftware, part of NGO-Opensource to enable SW technologies in the Metaportal project www.semanticwebsoftware.org -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: The next Internet giant: linking open data, providing open access to repositories
Sw-MetaPortal-ProjectParadigm wrote: Kinglsey, I would very much like to know what it will take to start up just such a federation. There are technical standards, legal issues of copyright and numerous others to explore. Federation post Web 2.0 will occur in the same fashion centralization occurred during the Web 1.0 and 2.0 stages of Web evolution i.e., services / solutions deployment models. If we take Web 2.0 as a example, the model is inherently central courtesy of a specific application of the SaaS (Software as a Service) deployment model. Google, Yahoo, Facebook, MySpace, and the litany of others (from Web 1.0 to 2.0) all fit my characterization in this regard. Federation will resurface by the emergence of personalized variations of the SaaS deployment model courtesy of cloud computing platform offerings like Amazon EC2 (and others). We will end up with Personal Portals that expose Metadata in RDF based Linked Data form (pun intended :-) ), which is something that our OpenLink Data Spaces platform (ODS) [1] has long facilitated (i.e., all your Web 2.0 silos are hooked into a virtual data space that exposes the data from those enclaves as RDF based Linked Data). How do we go about designing a framework for the federation, i.e. a defining framework document and the requirements for each participating space? See how we've done it via the ODS framework and programmers guide [2][3]. And to use analogies from the community of self-autonomous entities, will all spaces be equal like e.g. nations or will we have city-states, capitol districts, autonomous territories, provinces and the likes to identify the interrelatedness and dependency structure of the spaces. All of the above, in a nutshell. A federation presupposes common ideals, principles, standards, and intentions governed by a declaring, defining document and a constituting, instituting document. Well we are a loose-ish federation of social beings in our human meta-space, we just haven't translated that completely to cyberspace. I know of at least three spaces myself would like to join such a federation as soon as possible! I think all the silo-ed members of social-networks and other Web 2.0 solutions will make this happen in due course. The opportunity cost of futile resistance will force the silo owners to play ball (it always does eventually). We have all the essential building blocks in place, who will step to the fore to create and form part of a council of founding fathers for just such an endeavor? You [4] :-) The exciting part of this process is that both consumers groups of information and producers/providers of information represented in such a space of linked data could form part of the federation. By doing so you will create a much better feedback loop about what users would actually appreciate in functionality first, obviously building on the standards created by W3C et alii. An exciting idea this federation! Yes. Links: 1. http://virtuoso.openlinksw.com/wiki/main/Main/Ods 2. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/ODSProgrammersGuide 3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtuosoOdsUbiquityTutorials 4. http://tinyurl.com/6lwfqj Kingsley Milton Ponson GSM: +297 747 8280 Rainbow Warriors Core Foundation PO Box 1154, Oranjestad Aruba, Dutch Caribbean www.rainbowwarriors.net Project Paradigm: A structured approach to bringing the tools for sustainable development to all stakeholders worldwide www.projectparadigm.info NGO-Opensource: Creating ICT tools for NGOs worldwide for Project Paradigm www.ngo-opensource.org MetaPortal: providing online access to web sites and repositories of data and information for sustainable development www.metaportal.info SemanticWebSoftware, part of NGO-Opensource to enable SW technologies in the Metaportal project www.semanticwebsoftware.org --- On *Sun, 12/7/08, Kingsley Idehen /[EMAIL PROTECTED]/* wrote: From: Kingsley Idehen [EMAIL PROTECTED] Subject: Re: The next Internet giant: linking open data, providing open access to repositories To: [EMAIL PROTECTED] Cc: semantic-web [EMAIL PROTECTED], public-lod public-lod@w3.org Date: Sunday, December 7, 2008, 7:14 PM Sw-MetaPortal-ProjectParadigm wrote: The next Internet giant company will be linking open data and providing open access to repositories, in the process seamlessly combining both paid for subscriptions, Creative Commons or similar license based or open source software schemes. Revenues will be generated among other things from online advertising streams currently not utilized by Google or Yahoo! In the big scheme of things this company will redefine the concept of internet search to provide access to deep(er) web levels of data and information for which users will be willing to pay an annual flat fee subscription. Sound
Update: EC2 Editions of DBpedia 3.2 Neurocommons
All, DBpedia 3.2 [1] and Neurocommons [2] Linked Data Spaces now exist on EC2. In both cases you can now commission personal or service specific renditions of one or both data spaces. The time for commissioning each of these is 1.5 hrs (or less) compared to 8 -22 hrs from source for DBpedia and 1.15 hrs (or less) compared to 14 hrs or more for Neurocommons. Examples: 1. http://kingsley.idehen.name/resource/Linked_Data -- DBpedia 3.2 as a personal rendition on EC2 2. http://kingsley.idehen.name:8890 - Neurocommons in personal rendition form (see example HTML based Linked Data Space exploration at: http://kingsley.idehen.name:8890/science/owl/sciencecommons/gene_record) Installation Guide: 1. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIDBpediaInstall - DBpedia installation guide for EC2 2. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMINeuroCommonsInstall - Neurocommons installation guide for EC2 Next up, Bio2RDF :-) Note: The entire LOD data set collection is being constructed in parallel for similar deployment on EC2. Links: 1. http://dbpedia.org/About 2. http://neurocommons.org/page/Main_Page -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Commercial Product Announcements
Ian Davis wrote: Hi all, I'm not aware of any official policy on commercial posts to this list, but usual mailing list etiquette generally recommends that postings about commercial services and products should indicate their nature and give some indication of cost. I think this is even more important when the mailing list is focussed on a project with an open or free theme. Obviously, as I represent a vendor my preference is to allow commercial postings but I would also prefer that they are clearly labelled as such. Is this a policy that other members would like to see made explicit? Ian Ian there is no know policy for product announcements or job postings or anything else. I think we just collectively focus on on relevance. I think the relevance to Linked Data is simply what matters. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Commercial Product Announcements
Amy Stephen wrote: This could degenerate into a religious argument in short order. There is simply no way to specifically regulate conversation without throttling knowledge. Since this list is primarily focused on the enabling the free flow of knowledge, it seems a counterproductive move. If the list continues to focus on purpose and data and technologies and protocols and standards, and people try really, really hard to avoid self promotion of any variety as much as possible, and remember to be considerate of and trust one another, then things work. In the final analysis, cooperation and trust are our only hope for just about anything. Amen! Kingsley On Tue, Dec 9, 2008 at 6:25 AM, Kingsley Idehen [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote: David Wood wrote: On Dec 8, 2008, at 8:57 PM, Kingsley Idehen wrote: Ian Davis wrote: Hi all, I'm not aware of any official policy on commercial posts to this list, but usual mailing list etiquette generally recommends that postings about commercial services and products should indicate their nature and give some indication of cost. I think this is even more important when the mailing list is focussed on a project with an open or free theme. Obviously, as I represent a vendor my preference is to allow commercial postings but I would also prefer that they are clearly labelled as such. Is this a policy that other members would like to see made explicit? Ian Ian there is no know policy for product announcements or job postings or anything else. I think we just collectively focus on on relevance. I think the relevance to Linked Data is simply what matters. I agree, for what that is worth. Clear identification of commercial content is certainly appreciated, though. Even better would be identification capable of being filtered on. Regards, Dave Dave, So provide a template to clarify your suggestion. Ambiguity doesn't solve anything bar leaving grounds for subjective innuendo. Lets be as clear as possible since this is how objectivity is established. Also, for what it's worth, what is commercial content ? -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen http://www.openlinksw.com/blog/%7Ekidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Zemanta API: from text to LOD
Andraz Tori wrote: Hi, I am Andraz Tori, CTO at Zemanta. This is my first post to this list. I've read the discussion about commercial announcements just a day ago, so I hope this this announcement is relevant enough albeit it comes from a for-profit company. Welcome! Relevant! Please feel at home :-) I'd like to announce that Zemanta today launched a semantic API that is able to take plain text and disambiguate most important entities found to their meanings in Linking Open Data (currently dbPedia, MusicBrainz, Semantic CrunchBase and Freebase). I can't find the API itself or API demos. I've seen your illustrations etc.. As far as we know this is first such undertaking of this scale. Not necessarily, but being first doesn't matter, it's the value delivered that is ultimately the key factor (imho) :-) API is free to use up to 10k calls per day (default limit is 1k, email needs to be sent to raise it). Response can be in XML, JSON or RDF/XML. Developer info can be found at http://developer.zemanta.com and less technical explanations at http://www.zemanta.com/api I am very interested if this will make any new types of Mashups happen I would say potentially interesting meshups with the output of other Linked Data aware services, for instance. Everything meshes with everything in this realm once the essence is understood. Also I am interested in feedback on RDF/XML format/ontology. I see ourselves a bit like a stargate portal from unstructured text into the LOD, enabling LOD to be used in broader number of situations. Comments welcome :) Please send a link to a page that provides sample Linked Data example. Also clarify if you offer a REST style service as I only see your language specific APIs in your document web collective re. APIs etc.. I am assuming you offer something that works like this: 1. You receive a chunk of blurb as input 2. You analyze (where extraction and disambiguation occurs) 3. Return an RDF Linked Data graph in one or more formats as part of you output -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [Ann] OSM Linked Geo Data extraction, browser editor
On 12/10/08 1:25 PM, Sören Auer wrote: Hi all, We were working in the last weeks on bringing geo data derived from the marvelous OpenStreetMap project [1] to the data web. This work in progress is still far from being finished, however, we would like to share some first preliminary results: * A *vast amount of point-of-interest descriptions* was extracted from OSM and published as Linked Data at http://linkedgeodata.org * The *Linked Geo Data browser and editor* (available at http://linkedgeodata.org/browser) is a facet-based browser for geo content, which uses an OLAP inspired hypercube for quickly retrieving aggregated information about any user selected area on earth. Further information can be also found in our AKSW project description: http://aksw.org/Projects/LinkedGeoData Thanks go to Sebastian Dietzold, Jens Lehmann, Sebastian Hellmann, David Aumueller and other members of the AKSW team for their contributions. Merry Christmas to everybody from Leipzig Sören [1] http://openstreetmap.org Soren at Co., Very cool! -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: objects to facts to links to LOD to ???
On 12/16/08 10:44 AM, Knud Hinnerk Möller wrote: Hi, I'm not sure about the discovery part - that seems to be more a meta-layer with data describing linked datasets (VoiD [1]). However, I agree that inference would have to be in layer 4 and up. Things like non-explicit object consolidation, vocabulary and ontology mapping, etc. will be necessary to integrate the data even beyond the point of explicit links. @ravinder: I like the three layers you propose - is there a reference for that, or did you come up with it just now? I'd love to reference it in my PhD thesis! Cheers, Knud Knud et al, I think Ravinder has started the process of fixing the current Semantic Web layer cake :-) Which is a very good thing (imho, but not seeking a Layer Cake discussion explosion). The tricky part is the interchangeable nature of Discovery and Trust in any such scheme layer-wise. For instance, do Discovery and Trust occupy Layers 4, 5 or either ? We ultimately want to reason against trusted data sources, but the serendipity quotient of discovery is a key factor re. the dynamic nature of trusted sources. Since I am clearly thinking and writing (aloud) at the same time, I would suggest: Layer 4 - Discovery (with high Serendipity Quotient) Layer 5 - Trust (albeit inherently volatile) Kingsley [1] http://semanticweb.org/wiki/VoiD On 15.12.2008, at 20:54, Juan Sequeda wrote: Hi Ravinder, Interesting points. I would say that Layer 4 would be discovery. Furthermore, this would be an inference layer that would allow discovery. After having linked data, applications will be able to discover new data. So maybe layer 4 and up are part of the applications that enable discovery. Anyways, that is what I see what the semantic web is about: discovery and serendipity. Juan Sequeda, Ph.D Student Dept. of Computer Sciences The University of Texas at Austin www.juansequeda.com www.semanticwebaustin.org On Sat, Dec 13, 2008 at 11:16 AM, रविंदर ठाकुर (ravinder thakur) ravindertha...@gmail.com wrote: Hello Friends, We have now reasonably good repository of linked data but i am not able to find any good application being made out of it. From the dicsussion on these mailing lists and other places, i got the feeling that everyone in semantic world thinks that semantic web is something big but unfortunately nobody is able to think of an application for _general public_ that can be based on semantic web data we have currently. This led me to wonder that are there any more layers of information organization missing in the semantic web stack that are needed to generate good usable semantic web data which can be more useful for generating useful semantic apps . Few of the concepts i was thinking for upper layers to implement were like categorization, set formation(eg MIT passout in 2002), set intersection, etc. With these higher level layers I was hoping to build a system to find higher level relations say between places with low cancer rates and the food people eat there or any perticular gene found in such people. The current semantic stack looks like this: . layer5 (???) layer4 ( ???) layer3 linked facts (isa(mars,planet) AND partof(mars,solarsystem) AND mass(mars,6566.1255kg)) layer2 facts (isa(mars,planet), isa(cow,animal)) layer1 objects (eg cow, mars, man) Are there any other thoughts on the need of layers above the layer 3 (linked data ) or these layers will be defined by the respective apps developers ? Even if there isn't any need, i would atleast like to have a discussion on the kind upper level layers we might need :) thanks ravinder thakur - Knud Möller, MA +353 - 91 - 495086 Smile Group: http://smile.deri.ie Digital Enterprise Research Institute National University of Ireland, Galway Institiúid Taighde na Fiontraíochta Digití Ollscoil na hÉireann, Gaillimh -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Bio2Rdf EC2 AMI rendition is now live
All, Following along the lines of DBpedia [1] and NeuroCommons [2], we now have a rendition of Bio2Rdf.org [3] available on the Amazon EC2 Cloud. Total commissioning time: 1. 20 minutes for Virtuoso backup transfer from S3 bucket to EC2 AMI 2. 1.15 hrs for Virtuoso backup restoration. As per usual, there is a *temporary* instance deployed via one of my personal data spaces on EC2 which includes a SPARQL endpoint and HTML based linked data browser pages. Some sample HTML browser page links: 1. http://kingsley.idehen.name/geneid:4421752 2. http://kingsley.idehen.name/uniprot:A0A379 3. http://kingsley.idehen.name/citations:17309688 4. http://kingsley.idehen.name/isoforms:A0A379-1 Links: 1. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIDBpediaInstall 2. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMINeuroCommonsInstall 3. http://virtuoso.openlinksw.com/dataspace/dav/wiki/Main/VirtEC2AMIBio2rdfInstall -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
LOD Data Sets RDF Dump URLs
All, We are now ready to load the entire LOD cloud into a Virtuoso 6.0 (Cluster Edition) instance that will be available to the public (in identical manner to what we currently offer re. DBpedia). I would like everyone that has published an RDF data set archive, that seeks to have their data exposed by this instance, perform the following steps: 1. Go to: http://esw.w3.org/topic/DataSetRDFDumps 2. Add your data set to the table (if it isn't already listed) or correct erroneous entries 3. Add a URL entry to the Archive URL column 4. Add a Publisher URI to the Publisher / Maintainer column (* this URI will be used to construction Attribution Triples *) If you don't have a URI, you can get one very quickly by registering at: http://community.linkeddata.org/ods (*just register and you end up with a URI for yourself *). -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Hyperdata Hypercard
All, Please watch. http://www.archive.org/details/CC501_hypercard . Especially in relation to: http://www.youtube.com/watch?v=hAAwKxeoCrk History is always a great tutor when journeying into the future. Concepts are quite static relative to contextual application and comprehension in an innovation continuum. The Web is an innovation continuum that provides new and improved context for applying and comprehending old concepts. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
CORRECTION: Hyperdata Hypercard
On 12/22/08 5:40 PM, Kingsley Idehen wrote: All, Please watch. http://www.archive.org/details/CC501_hypercard . Especially in relation to: http://www.youtube.com/watch?v=hAAwKxeoCrk History is always a great tutor when journeying into the future. Concepts are quite static relative to contextual application and comprehension in an innovation continuum. The Web is an innovation continuum that provides new and improved context for applying and comprehending old concepts. The Web is part of an innovation continuum that provides new and improved context for applying and comprehending old concepts. was what I meant to say :-) -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Granular dereferencing ( prop by prop ) using REST + LinkedData; Ideas?
On 12/29/08 10:51 AM, Aldo Bucchi wrote: Hi All, I am in the process of LODing a dataset in which certain properties are generated on the fly ( props derived from aggregate calculations over the dataset, remote calls, etc ). I would like to let the clients choose which of these expensive properties they need on demand and on a granular level. For example, lets say I am interested in knowing more about resource http://ex.com/a. Per LD conventions, dereferencing http://ex.com/a ( via 303 ) returns http://ex.com/a a ex:Thing ; rdfs:label a sample dynamic resource; ex:dynamic1 45567 . The problem is that the value for ex:dynamic1 is very expensive to compute. Therefore, I would like to partition the document in such a way that the client can ask for the property on a lazy, deferred manner ( a second call in the future ). The same is true for dynamic2, dynamic3, dynamic4, etc. All should be retrievable independently and on demand. * I am aware that this can be achieved by extending SPARQL in some toolkits. But I need LOD. * I am also aware that most solutions require us to break URI obscurity by stuffing the subject and predicate in the uri for a doc. * Finally, seeAlso is too broad as it doesn't convey the information I need. Anyone came up with a clean pattern for this? Ideas? Something as simple as: GET http://x.com/sp?s={subject}p={predicate} --- required RDF works for me... but... . If possible, I would like to break conventions in a conventional manner ;) Best, A Aldo, We will have something to demonstrate re. this matter once we are done with the Xmas / New Year break in relation to the entire LOD data set being loaded into Virtuoso. In a nutshell, we are going to expose Quad Store stats using VoiD. The Client shouldn't have to calculate anything anymore, the stats should be part of the data exposed by the RDF store. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Granular dereferencing ( prop by prop ) using REST + LinkedData; Ideas?
On 1/4/09 9:28 AM, Yves Raimond wrote: Hi Richard! On 3 Jan 2009, at 12:23, Richard Cyganiak rich...@cyganiak.de wrote: On 2 Jan 2009, at 23:20, Yves Raimond wrote: snip I proposed this solution: http://simile.mit.edu/mail/ReadMsg?listName=Linking%20Open%20DatamsgId=20926 And some refinements here: http://simile.mit.edu/mail/ReadMsg?listName=Linking%20Open%20DatamsgId=20962 snip This discussion was indeed interesting :) I now tend to think that linking to a separate document is a cleaner way to go, but I am still concerned about auto-discovery. When you see something like: :New_York :morePersonsBornHere http://example.org/persons_nyc.rdf . in the representation of :New_York, you still need a way to describe the fact that :morePersonsBornHere links to a document holding lots of :birthPlace properties. Just saying that :morePersonsBornHere rdfs:subPropertyOf rdfs:seeAlso won't do the job properly - how can you tell before retrieving the whole document? Yves, the proposal above addresses this. There would be a triple: :birthPlace link:subjectListProperty :morePersonsBornHere . This triple can be either directly in the :New_York description, or in the vocabulary (where you'd find it by dereferencing :morePersonsBornHere). The triple tells clients that they should follow the :morePersonsBornHere link if they are interested in :birthPlace triples. So, autodiscovery is solved. Yes, it does work, but only for simple property lists. What about find here persons born in NYC between 1945 and 1975 ? But perhaps the approach I proposed when we discussed the void:example property could work, in exactly the same way as in [1]. In the representation of :New_York, we could write something like (in N3): http://example.org/persons_nyc.rdf void:example { :al_pacino :birthPlace :New_York }. N3 formulae cannot be expressed in RDFa or RDF/XML. How would you serialize this in practice? As in the post I refered to: you can point to http://example.org/dataset-example.rdf where you put these example triples. Then, a SPARQL query like the following could find the documents that hold information about persons born in New York: SELECT ?doc WHERE { ?doc void:example ?g . GRAPH ?g { ?person :birthPlace :New_York . } } One of the good thing with this approach is that the patterns of information held in the target document can be arbitrarily complex - As far as I can remember, all the examples that people have given could be addressed with a simple property-based approach. Has anyone mentioned a use case that goes beyond looking for a single property? If not, then what does the additional complexity of this proposal buy us in practice? The example mentioned in my post uses more than one property, or the exampl above. “This is good because it can deal with arbitrary complexity” is a fallacy on the Web. Usually it translates as: “This is bad because it makes the common case harder.” (I note that the situation here is different from what you described in [1]. There it was about annotations on a dataset level. Here it is about annotating links that occur within many or all individual documents of a dataset.) A RDF document is a dataset, and can be described as such :-) Yves, Just to clarify language (I do agree with your point). An RDF document is a resource (of type/kind information resource) and as a result -- like all Things -- it can be formally described via its attributes and relationship properties :-) Happy New Year! Kingsley Cheers! y Best, Richard and the only thing you have to do is to provide an example RDF graph, holding something representative of what you put in that document. Cheers! y [1] http://blog.dbtune.org/post/2008/06/12/Describing-the-content-of-RDF-datasets -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: Contd: UMBEL Effect re. Faceted Browsing
On 1/9/09 4:32 PM, Jason Kolb wrote: That's cool stuff Kingsley. Seeing the two next to each other (with and w/out Umbel) makes it easy to see the value that Umbel adds. Jason, Yes, which is why I took an oath (with myself) last year to only demo or write about UMBEL when I have a simple high impact demo :-) More to come. Happy New Year! Kingsley On Fri, Jan 9, 2009 at 4:25 PM, Kingsley Idehen kide...@openlinksw.com mailto:kide...@openlinksw.com wrote: All, One thing that should emerge from this server side faceted browser demo is the impact of inferencing based on UMBEL ontology. Remember, UMBEL provides binding between the OpenCyc Upper Ontology and domain specific ontologies such as FOAF. Compare items 12 for an example of the effect of this Ontology Meshup: 1. http://linkeddata.uriburner.com/about/html/http://xmlns.com/foaf/0.1/Person (no UMBEL inferencing enabled) 2. http://b3s.openlinksw.com/about/?url=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2FPersonsid=177 http://b3s.openlinksw.com/about/?url=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2FPersonsid=177 (UMBEL inferencing enabled) Note the expanse of the FOAF vocabulary's graph courtesy of UMBEL linkage to other ontologies via Equivalence, Subclass, and other links (what Fred refers to as domain explosion in some of his blog posts). -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen http://www.openlinksw.com/blog/%7Ekidehen President CEO OpenLink Software Web: http://www.openlinksw.com -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: [ANN] t4gm.info: a Linked Data rendering of the Thesaurus of Graphic Materials
On 1/14/09 6:13 AM, Bradley Allen wrote: t4gm.info http://www.t4gm.info is a simple, lightweight and DRY http://en.wikipedia.org/wiki/DRY Linked Open Data rendering in XHTML+RDFa of the Library of Congress http://www.loc.gov/index.html' Thesaurus for Graphic Materials http://lcweb2.loc.gov/pp/tgmhtml/tgmabt.html. Details about its use and implementation can be found on its help page http://www.t4gm.info/help. Comments to the undersigned are welcomed. - regards, BPA Bradley P. Allen http://bradleypallen.org +1 310 951 4300 Bradley, Very cool!! -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: RKBExplorer Datasets in LOD Cloud Diagram (was Re: FW: [ESW Wiki] Update of SweoIG/TaskForces/CommunityProjects/LinkingOpenData by TedThibodeauJr)
and delete it. Any use of this e-mail by an unauthorised recipient is prohibited. Talis Information Ltd is a member of the Talis Group of companies and is registered in England No 3638278 with its registered office at Knights Court, Solihull Parkway, Birmingham Business Park, B37 7YB. __ This email has been scanned by the MessageLabs Email Security System. For more information please visit http://www.messagelabs.com/email __ -- Dr Tom Heath Researcher Platform Division Talis Information Ltd T: 0870 400 5000 W: http://www.talis.com/ -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
DBpedia Update and Pending Changeover
All, We are about to transition DBpedia from a Virtuoso 5 instance over to a Virtuoso 6.x (Cluster Edition) instance sometime this week after a few days of live testing against a temporary secondary instance [1]. The new release also includes our much vaunted server side faceted browsing [2] that leverages SPARQL aggregates and configurable interaction time for query payloads (*note the retry feature which increments the interactive time window to demonstrate full effects*). In the process of conducting this effort, we discovered that the Yago Class Hierarchy hadn't been loaded (courtesy of some TBox subsumption work performed by the W3C's HCLS project participants such as Rob Frost). In addition, Virtuoso 6.x introduces TBox subsumption via new Virtuoso-SPARQL transitivity options as exemplified below using the V6.0 instance at: http://dbpedia2.openlinksw.com:8895/sparql (or /isparql). -- Yago Subsumption based AlphaReceptors query select ?y from http://dbpedia.org/resource/classes/yago# where { { select * where { ?x rdfs:subClassOf ?y . } } option (transitive, t_distinct, t_in (?x), t_out (?y) ) . filter (?x = http://dbpedia.org/class/yago/AlphaReceptor105609111) } -- Yago Subsumption based Receptors query select ?x from http://dbpedia.org/resource/classes/yago# where { { select * where { ?x rdfs:subClassOf ?y . } } option (transitive, t_distinct, t_in (?x), t_out (?y) ) . filter (?y = http://dbpedia.org/class/yago/Receptor105608868) } Links: 1. http://dbpedia.org/resource/DBpedia - test URI 2. http://dbpedia2.openlinksw.com:8895/fct/facet.vsp - server hosted faceted browser. -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com
Re: studies about LD visibility
On 1/27/09 1:24 PM, Samuel Rose wrote: Things that cannot be done without LD: I don't think there is technically anything that explicitly cannot be done without Linking Open Data. Samuel, I profoundly believe that there are things that cannot be achieved without Linked Data. Now before I progress, let be clear about Linked Data. We have: 1. Linked Data the meme from TimBL re. adding datum level linkage between resources on the Web (hyperdata) which is (imho) short for: Linked Data Web or Web of Linked Data 2. Linked Data in a much broader sense covering the application / incorporation of HTTP into the time-tested Data Access By Reference pattern (pointers) commonly used at the OS and DBMS levels to access and manipulate data. Irrespective of 12 specifics, there is one common theme: integration of disparate data sources without application, operating system, or network level impediments. Basically, untethered data access by reference courtesy of an HTTP based pointer system. Note: The ubiquity of HTTP, combined with its inherent ability to negotiate referent (what you point to with a URI) description representation, is the source of its potency. For eons (computer industry time) in the distributed computing realm, the ability to reference and de-reference an object (datum) via a pointer that's functional across application, operating system, and network boundaries has been the holy grail ; many technologies have tried an failed re. this pursuit, and the stories are engraved in the technology tombstones of efforts such as: corba, cairo, copeland, opendoc, ole, and others. I hope I've clarified what Linked Data's USP really is. btw - Linking Open Data (LOD) is about a community that adheres to a set of best practices for publishing data on the Web in a manner that facilitate datum level linkage (hyperdata links) in addition to existing document level linkage (hypertext links). -- Regards, Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen President CEO OpenLink Software Web: http://www.openlinksw.com