Re: linked data about the W3C?
Pierre-Antoine Champin, You mean something like [1]? Cheers, Michael [1] http://www.w3.org/2002/01/tr-automation/tr.rdf -- Dr. Michael Hausenblas, Research Fellow LiDRC - Linked Data Research Centre DERI - Digital Enterprise Research Institute NUIG - National University of Ireland, Galway Ireland, Europe Tel. +353 91 495730 http://linkeddata.deri.ie/ http://sw-app.org/about.html From: Pierre-Antoine Champin swlists-040...@champin.net Date: Wed, 24 Nov 2010 12:40:33 +0100 To: Linked Data community public-lod@w3.org Subject: linked data about the W3C? Resent-From: Linked Data community public-lod@w3.org Resent-Date: Wed, 24 Nov 2010 11:42:38 + Hi, is there any linked data published about the W3C? Semantic Radar does not see any on their webpages. It is a shame, because I would like to get some statistics about the members, and I'm condemned to manually go through the 324 links of [1]. Talking about eating one's own dog food :-P pa [1] http://www.w3.org/Consortium/Member/List
FW: Failed to port datastore to RDF, will go Mongo
Friedrich, I'm forwarding your message to one of the W3 lists. Some of your questions could be easily answered (e.g. for euro in your context, you don't have a predicate for that, you have an Observation with units of a currency and you could take the currency from dbpedia, the predicate is units). But I think your concerns are quite valid generally and your experience reflects that of most web site developers that encounter RDF. LOD list, Friedrich is a clueful developer, responsible for http://bund.offenerhaushalt.de/ amongst other things. What can we learn from this? How do we make this better? -w - Forwarded message from Friedrich Lindenberg friedr...@pudo.org - From: Friedrich Lindenberg friedr...@pudo.org Date: Wed, 24 Nov 2010 11:56:20 +0100 Message-Id: a9089567-6107-4b43-b442-d09dcc0c3...@pudo.org To: wdmmg-discuss wdmmg-disc...@lists.okfn.org Subject: [wdmmg-discuss] Failed to port datastore to RDF, will go Mongo (reposting to list): Hi all, As an action from OGDCamp, Rufus and I agreed that we should resume porting WDMMG to RDF in order to make the data model more flexible and to allow a merger between WDMMG, OffenerHaushalt and similar other projects. After a few days, I'm now over the whole idea of porting WDMMG to RDF. Having written a long technical pro/con email before (that I assume contained nothing you don't already know), I think the net effect of using RDF would be the following: * Lots of coolness, sucking up to linked data people. * Further research regarding knowledge representation. vs. * Unstable and outdated technological base. No triplestore I have seen so far seemed on par with MySQL 4. * No freedom wrt to schema, instead modelling overhead. Spent 30 minutes trying to find a predicate for Euro. * Scares off developers. Invested 2 days researching this, which is how long it took me to implement OHs backend the first time around. Project would need to be sustained through linked data grad students. * Less flexibility wrt to analytics, querying and aggregation. SPARQL not so hot. * Good chance of chewing up the UI, much harder to implement editing. I normally enjoy learning new stuff. This is just painful. Most of the above points are probably based on my ignorance, but it really shouldn't take a PhD to process some gov spending tables. I'll now start a mongo effort because I really think this should go schema-free + I want to get stuff moving. If you can hold off loading Uganda and Israel for a week that would of course be very cool, we could then try to evaluate how far this went. Progress will be at: http://bitbucket.org/pudo/wdmmg-core Friedrich ___ wdmmg-discuss mailing list wdmmg-disc...@lists.okfn.org http://lists.okfn.org/mailman/listinfo/wdmmg-discuss - End forwarded message - -- William Waites http://eris.okfn.org/ww/foaf#i 9C7E F636 52F6 1004 E40A E565 98E3 BBF3 8320 7664
Re: FW: Failed to port datastore to RDF, will go Mongo
... on the plus side, Friedrich wrote: ] * Lots of coolness, sucking up to linked data people. I don't see these as particularly good things in themselves. The solutions have to be obviously technically sound and convenient to use. Drinking the kool-aid is not helpful. * [2010-11-24 08:05:08 -0500] Kingsley Idehen kide...@openlinksw.com écrit: ] ] Is your data available as a dump? UK data for 2009 that I made is available at: http://semantic.ckan.net/dataset/cra/2009/dump.nt.bz2 http://semantic.ckan.net/dataset/cra/2009/dump.nq.bz2 But this was done more or less by hand and repurposing the CSV - SDMX (this was done before QB became best practice) scripts is not easy. Still, from a modeling perspective they might be a good starting point. But having to ask a question in the right place and the answer being a good starting point is maybe different from doing a google search and finding easy to follow recipes that can immediately plugged into some web app. Cheers, -w -- William Waites http://eris.okfn.org/ww/foaf#i 9C7E F636 52F6 1004 E40A E565 98E3 BBF3 8320 7664
Re: FW: Failed to port datastore to RDF, will go Mongo
Hi William, Friederich. This is an excellent email. My replies inlined. Hope I can help. On Wed, Nov 24, 2010 at 9:47 AM, William Waites w...@styx.org wrote: Friedrich, I'm forwarding your message to one of the W3 lists. Some of your questions could be easily answered (e.g. for euro in your context, you don't have a predicate for that, you have an Observation with units of a currency and you could take the currency from dbpedia, the predicate is units). But I think your concerns are quite valid generally and your experience reflects that of most web site developers that encounter RDF. LOD list, Friedrich is a clueful developer, responsible for http://bund.offenerhaushalt.de/ amongst other things. What can we learn from this? How do we make this better? -w - Forwarded message from Friedrich Lindenberg friedr...@pudo.org - From: Friedrich Lindenberg friedr...@pudo.org Date: Wed, 24 Nov 2010 11:56:20 +0100 Message-Id: a9089567-6107-4b43-b442-d09dcc0c3...@pudo.org To: wdmmg-discuss wdmmg-disc...@lists.okfn.org Subject: [wdmmg-discuss] Failed to port datastore to RDF, will go Mongo (reposting to list): Hi all, As an action from OGDCamp, Rufus and I agreed that we should resume porting WDMMG to RDF in order to make the data model more flexible and to allow a merger between WDMMG, OffenerHaushalt and similar other projects. After a few days, I'm now over the whole idea of porting WDMMG to RDF. Having written a long technical pro/con email before (that I assume contained nothing you don't already know), I think the net effect of using RDF would be the following: * Lots of coolness, sucking up to linked data people. * Further research regarding knowledge representation. I will quickly outline some points that I think are advantages from a developer POV. ( once you tackle the problems you outline below, of course ). * A highly expressive language ( SPARQL ) * Ease of creating workflows where data moves from one app to another. And this is not just buzz. The self-contained nature of triples and IDs make it so that you can SPARQL select on one side and SPARQL insert on another. I do this all the time, creating data pipelines. I admit it has taken some time to master, but I can peform magic from my customer's point of view. vs. * Unstable and outdated technological base. No triplestore I have seen so far seemed on par with MySQL 4. * You definitely need to give Virtuoso a try. It is a mature SQL database that grew into RDF. I Strongly disagree with this point as I have personally created highly demanding projects for large companies using Virtuoso's Quad Store. To give you a real life case, the recent Brazilian Election portal by Globo.com ( http://g1.globo.com/especiais/eleicoes-2010/ ) has Virtuoso under the hood and, being a highly important, mission critical app in a major ( 4th ) media company it is not a toy application. I know many others but in this one I participated so I can tell you it is Virtuoso w/o fear mistake. * No freedom wrt to schema, instead modelling overhead. Spent 30 minutes trying to find a predicate for Euro. Yes! This is a major problem and we as a community need to tackle it. I am intrigued to see what ideas come up in this thread. Thanks for bringing it up. As an alternative, you can initially model everything using a simple urn:foo:xxx or http://mydomain.com/id/xxx schema ( this is what I do ) and as you move fwd you can refactor the model. Or not. You can leave it as is and it will still be integratable ( able to live along other datasets in the same store ). Deploying the Linked part of Linked Data ( the dereferencing protocols ) later on is another game. * Scares off developers. Invested 2 days researching this, which is how long it took me to implement OHs backend the first time around. Project would need to be sustained through linked data grad students. * Less flexibility wrt to analytics, querying and aggregation. SPARQL not so hot. Did you try Virtuoso? Seriously. It provides out of the box common aggregates and is highly extensible. You basically have a development platform at your disposal. * Good chance of chewing up the UI, much harder to implement editing. Definitely hard. This is something I hope will be alleviated once we start getting more demos into the wild. But, take note: the Active Record + MVC pattern works. This is not as alien as it seems. Also, SPARQL also removes the joines as some of the major NoSQL offerings do. I find it terribly easy to create UIs over RDF, but I have been doing it for a while already. I normally enjoy learning new stuff. This is just painful. Most of the above points are probably based on my ignorance, but it really shouldn't take a PhD to process some gov spending tables. I'll now start a mongo effort because I really think this should go schema-free + I want to get stuff moving. If you can hold off loading Uganda and Israel for a week that
Re: FW: Failed to port datastore to RDF, will go Mongo
Sorry, I forgot to add something critical. Ease of integration ( moving triples ) is just the beginning. Once you get a hold on the power of ontologies and inference as views your data starts becoming more and more useful. But the first step is getting your data into RDF and the return on that investment is SPARQL and the ease to integrate. I usually end up with several transformation pipelines and accesory TTL files which get all combined into one dataset. TTLs are easily editable by hand, collaboratively versiones, while giving you full expressivity. TTL files alone are why some developers fall in love with Linked Data. On Wed, Nov 24, 2010 at 10:33 AM, Aldo Bucchi aldo.buc...@gmail.com wrote: Hi William, Friederich. This is an excellent email. My replies inlined. Hope I can help. On Wed, Nov 24, 2010 at 9:47 AM, William Waites w...@styx.org wrote: Friedrich, I'm forwarding your message to one of the W3 lists. Some of your questions could be easily answered (e.g. for euro in your context, you don't have a predicate for that, you have an Observation with units of a currency and you could take the currency from dbpedia, the predicate is units). But I think your concerns are quite valid generally and your experience reflects that of most web site developers that encounter RDF. LOD list, Friedrich is a clueful developer, responsible for http://bund.offenerhaushalt.de/ amongst other things. What can we learn from this? How do we make this better? -w - Forwarded message from Friedrich Lindenberg friedr...@pudo.org - From: Friedrich Lindenberg friedr...@pudo.org Date: Wed, 24 Nov 2010 11:56:20 +0100 Message-Id: a9089567-6107-4b43-b442-d09dcc0c3...@pudo.org To: wdmmg-discuss wdmmg-disc...@lists.okfn.org Subject: [wdmmg-discuss] Failed to port datastore to RDF, will go Mongo (reposting to list): Hi all, As an action from OGDCamp, Rufus and I agreed that we should resume porting WDMMG to RDF in order to make the data model more flexible and to allow a merger between WDMMG, OffenerHaushalt and similar other projects. After a few days, I'm now over the whole idea of porting WDMMG to RDF. Having written a long technical pro/con email before (that I assume contained nothing you don't already know), I think the net effect of using RDF would be the following: * Lots of coolness, sucking up to linked data people. * Further research regarding knowledge representation. I will quickly outline some points that I think are advantages from a developer POV. ( once you tackle the problems you outline below, of course ). * A highly expressive language ( SPARQL ) * Ease of creating workflows where data moves from one app to another. And this is not just buzz. The self-contained nature of triples and IDs make it so that you can SPARQL select on one side and SPARQL insert on another. I do this all the time, creating data pipelines. I admit it has taken some time to master, but I can peform magic from my customer's point of view. vs. * Unstable and outdated technological base. No triplestore I have seen so far seemed on par with MySQL 4. * You definitely need to give Virtuoso a try. It is a mature SQL database that grew into RDF. I Strongly disagree with this point as I have personally created highly demanding projects for large companies using Virtuoso's Quad Store. To give you a real life case, the recent Brazilian Election portal by Globo.com ( http://g1.globo.com/especiais/eleicoes-2010/ ) has Virtuoso under the hood and, being a highly important, mission critical app in a major ( 4th ) media company it is not a toy application. I know many others but in this one I participated so I can tell you it is Virtuoso w/o fear mistake. * No freedom wrt to schema, instead modelling overhead. Spent 30 minutes trying to find a predicate for Euro. Yes! This is a major problem and we as a community need to tackle it. I am intrigued to see what ideas come up in this thread. Thanks for bringing it up. As an alternative, you can initially model everything using a simple urn:foo:xxx or http://mydomain.com/id/xxx schema ( this is what I do ) and as you move fwd you can refactor the model. Or not. You can leave it as is and it will still be integratable ( able to live along other datasets in the same store ). Deploying the Linked part of Linked Data ( the dereferencing protocols ) later on is another game. * Scares off developers. Invested 2 days researching this, which is how long it took me to implement OHs backend the first time around. Project would need to be sustained through linked data grad students. * Less flexibility wrt to analytics, querying and aggregation. SPARQL not so hot. Did you try Virtuoso? Seriously. It provides out of the box common aggregates and is highly extensible. You basically have a development platform at your disposal. * Good chance of chewing up the UI, much harder to implement editing.
Re: linked data about the W3C?
Agreed. I'm long thinking that working group activities (people, emails, meetings, actions) should be queryable via triplification. W3C's various semantic wikis should also be better used for data building. Jie On Wed, Nov 24, 2010 at 06:40, Pierre-Antoine Champin swlists-040...@champin.net wrote: Hi, is there any linked data published about the W3C? Semantic Radar does not see any on their webpages. It is a shame, because I would like to get some statistics about the members, and I'm condemned to manually go through the 324 links of [1]. Talking about eating one's own dog food :-P pa [1] http://www.w3.org/Consortium/Member/List
Re: linked data about the W3C?
Le 24 nov. 2010 à 06:40, Pierre-Antoine Champin a écrit : is there any linked data published about the W3C? Semantic Radar does not see any on their webpages. QA information on specifications http://www.w3.org/QA/TheMatrix.rdf List of all Talks of W3C http://www.w3.org/2007/08/pyRdfa/extract?uri=http://www.w3.org/2007/11/Talks/all.html Recs http://www.w3.org/2003/03/recs.rdf http://www.w3.org/2002/01/tr-automation/tr.rdf A few of them. There are plenty of things around Minutes of meetings etc. http://www.google.com/search?q=filetype%3Ardf+site%3Aw3.org -- Karl Dubost - http://dev.opera.com/ Developer Relations Tools, Opera Software
Re: FW: Failed to port datastore to RDF, will go Mongo
On 11/24/10 8:29 AM, William Waites wrote: ... on the plus side, Friedrich wrote: ] * Lots of coolness, sucking up to linked data people. I don't see these as particularly good things in themselves. The solutions have to be obviously technically sound and convenient to use. Drinking the kool-aid is not helpful. * [2010-11-24 08:05:08 -0500] Kingsley Idehenkide...@openlinksw.com écrit: ] ] Is your data available as a dump? UK data for 2009 that I made is available at: http://semantic.ckan.net/dataset/cra/2009/dump.nt.bz2 http://semantic.ckan.net/dataset/cra/2009/dump.nq.bz2 But this was done more or less by hand and repurposing the CSV - SDMX (this was done before QB became best practice) scripts is not easy. Still, from a modeling perspective they might be a good starting point. But having to ask a question in the right place and the answer being a good starting point is maybe different from doing a google search and finding easy to follow recipes that can immediately plugged into some web app. Cheers, -w William, What does MySQL 4 do with this data that can't be done with a moderately capable RDF quad / triplestore? If I am going to run rings around this thing, I need a starting point :-) -- Regards, Kingsley Idehen President CEO OpenLink Software Web: http://www.openlinksw.com Weblog: http://www.openlinksw.com/blog/~kidehen Twitter/Identi.ca: kidehen
Re: FW: Failed to port datastore to RDF, will go Mongo
On Wed, 2010-11-24 at 12:51 -0500, Kingsley Idehen wrote: What does MySQL 4 do with this data that can't be done with a moderately capable RDF quad / triplestore? If I am going to run rings around this thing, I need a starting point :-) That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. This is about the 1000th time I have heard this story, and the ability range of those saying the same thing is huge - from 9-5 devs who learn what they need to people who research and teach artificial intelligence and other cutting edge areas and who actively learn new, complex skills just because they can. The point is not whether someone who (co?)developed the virtuoso triplestore can make RDF work, it's whether someone with the time, current documentation and inclination can. Ben -- Regards, Kingsley Idehen President CEO OpenLink Software Web: http://www.openlinksw.com Weblog: http://www.openlinksw.com/blog/~kidehen Twitter/Identi.ca: kidehen
Re: linked data about the W3C?
Aye, young man we have had linked data about W3C for many years :-) Point tabulator at http://www.3.org/data#W3C That wil pull in the organizational structure and the specs: - domains - activities - groups - deliverables - documents, versions etc all linked. Not sure how up to date. It does not currently include membership. I wonder what it would take to get it on the Semantic Radar of which your speak. I have added link rel=alternate type=application/rdf+xml title=Linked data about W3C href=data / to the home page (revision 4.2913). Maybe that will do it. Tim BL In http://lists.w3.org/Archives/Public/public-lod/2010Nov/0521.html Pierre-Antoine Champin swlists-040...@champin.net wrote: Hi, is there any linked data published about the W3C? Semantic Radar does not see any on their webpages. It is a shame, because I would like to get some statistics about the members, and I'm condemned to manually go through the 324 links of [1]. Talking about eating one's own dog food :-P pa
Re: linked data about the W3C?
Le 24 nov. 2010 à 13:15, Tim Berners-Lee a écrit : Aye, young man we have had linked data about W3C for many years :-) Point tabulator at http://www.3.org/data#W3C http://www.w3.org/data#W3C sir ;) -- Karl Dubost - http://dev.opera.com/ Developer Relations Tools, Opera Software
Re: FW: Failed to port datastore to RDF, will go Mongo
Ben, Just like we can 1) download xampp and install php, apache, mysql with one click 2) open a browser, open phpmyadmin, create my db 3) copy paste any snippet of code I can find on the web about connecting php/java etc to a mysql database 4) write code to select/insert/update my db ... you are asking for these same 4 simple steps but for an RDF database? Juan Sequeda +1-575-SEQ-UEDA www.juansequeda.com On Wed, Nov 24, 2010 at 12:12 PM, Ben O'Steen bost...@gmail.com wrote: On Wed, 2010-11-24 at 12:51 -0500, Kingsley Idehen wrote: What does MySQL 4 do with this data that can't be done with a moderately capable RDF quad / triplestore? If I am going to run rings around this thing, I need a starting point :-) That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. This is about the 1000th time I have heard this story, and the ability range of those saying the same thing is huge - from 9-5 devs who learn what they need to people who research and teach artificial intelligence and other cutting edge areas and who actively learn new, complex skills just because they can. The point is not whether someone who (co?)developed the virtuoso triplestore can make RDF work, it's whether someone with the time, current documentation and inclination can. Ben -- Regards, Kingsley Idehen President CEO OpenLink Software Web: http://www.openlinksw.com Weblog: http://www.openlinksw.com/blog/~kidehen Twitter/Identi.ca: kidehen
Re: FW: Failed to port datastore to RDF, will go Mongo
On Wed, 2010-11-24 at 14:40 -0600, Juan Sequeda wrote: Ben, Just like we can 1) download xampp and install php, apache, mysql with one click 2) open a browser, open phpmyadmin, create my db 3) copy paste any snippet of code I can find on the web about connecting php/java etc to a mysql database 4) write code to select/insert/update my db ... you are asking for these same 4 simple steps but for an RDF database? Not me personally, but in my experience of talking to developers in the HE/FE sector as well as commercial devs through JISC, running Dev8D and so on, being able to achieve those steps in the manner you have suggested is crucial. Yes, I exaggerated about my hearing the same tale a thousand times, but I have heard that perception of RDF/triplestores many, many times as unfounded as some may argue it is. This will sound like heresy, but the closest parallel I've found to step 1) is with mulgara (excepting that a Java runtime of some sort is required.) Run the jar, open browser, and run through the web-based examples that cover input, update and query. Ben Juan Sequeda +1-575-SEQ-UEDA www.juansequeda.com On Wed, Nov 24, 2010 at 12:12 PM, Ben O'Steen bost...@gmail.com wrote: On Wed, 2010-11-24 at 12:51 -0500, Kingsley Idehen wrote: What does MySQL 4 do with this data that can't be done with a moderately capable RDF quad / triplestore? If I am going to run rings around this thing, I need a starting point :-) That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. This is about the 1000th time I have heard this story, and the ability range of those saying the same thing is huge - from 9-5 devs who learn what they need to people who research and teach artificial intelligence and other cutting edge areas and who actively learn new, complex skills just because they can. The point is not whether someone who (co?)developed the virtuoso triplestore can make RDF work, it's whether someone with the time, current documentation and inclination can. Ben -- Regards, Kingsley Idehen President CEO OpenLink Software Web: http://www.openlinksw.com Weblog: http://www.openlinksw.com/blog/~kidehen Twitter/Identi.ca: kidehen
Re: FW: Failed to port datastore to RDF, will go Mongo
On 11/24/10 3:57 PM, Ben O'Steen wrote: On Wed, 2010-11-24 at 14:40 -0600, Juan Sequeda wrote: Ben, Just like we can 1) download xampp and install php, apache, mysql with one click 2) open a browser, open phpmyadmin, create my db 3) copy paste any snippet of code I can find on the web about connecting php/java etc to a mysql database 4) write code to select/insert/update my db ... you are asking for these same 4 simple steps but for an RDF database? Not me personally, but in my experience of talking to developers in the HE/FE sector as well as commercial devs through JISC, running Dev8D and so on, being able to achieve those steps in the manner you have suggested is crucial. Yes, I exaggerated about my hearing the same tale a thousand times, but I have heard that perception of RDF/triplestores many, many times as unfounded as some may argue it is. This will sound like heresy, but the closest parallel I've found to step 1) is with mulgara (excepting that a Java runtime of some sort is required.) Run the jar, open browser, and run through the web-based examples that cover input, update and query. Ben You should make a basic RDBMS your yardstick e.g. FoxPRO, Access, Filemaker, SQL Server etc.. If you have enterprise DBMS experience then: Oracle, SQL Server, Sybase, Ingres, Informix, Progress (OpenEdge), Firebird (once Interbase), PostgreSQL, MySQL. In all cases, this is what developers do: 1. Install DBMS 2. Load (using various data loaders and import utilities) and Create Data 3. Create Views and Queries 4. Put Forms and Reports atop Views (or Tables) 5. Enjoy power of RDBMS apps. This is what end-users (basic or power users do): 1. Get a productivity tool of choice (Word Processor, Spreadsheet, Report Writer etc) 2. Connect to RDBMS via an ODBC Data Source Name (which via ODBC Driver Manager is bound to Drivers for each DBMS) 3. Enjoys power of RDBMS via their preferred Desktop tool. Here is what so called Web Developers do: 1. Find an Open Source DBMS 2. Compile it 3. Work through LAMP stack to PHP, Pyton, Ruby, TCL, others 4. Ignore DBMS independent API of ODBC (available via iODBC or unixODBC effort) and couple HTML pages directly to DBMS 5. Ignore DBMS for user account management and stick that in HTML page layer too. Irrespective of where you fit in re. the above, this is what you should be able to do with Relational Property Graph Databases that support resolvable URIs as Unique Keys: 1. Load data - there are a myriad of paths including transient and materialized views over ODBC or JDBC accessible RDBMS data sources, Web Services, many other data container formats (spreadsheets, CSV files, etc..) 2. Use HTML+RDFa (or basic HTML) pages as Forms and Report Writer tool re. data browsing 3. Enjoy power of Linked Data. Note re. above: 1. No re-write rules coding 2. No 303 debates re. how to make Unique Keys resolve 3. No exposure to Name | Address disambiguation re. Unique Keys . Mulgara already mandates Java. Java != Platform Independent either. I am mandating nothing bar installing a DBMS and then simply leveraging HTTP, EAV Data Model, and the power of a Relational Property Graph Database that may or may not provide output in RDF format. Kingsley Juan Sequeda +1-575-SEQ-UEDA www.juansequeda.com On Wed, Nov 24, 2010 at 12:12 PM, Ben O'Steenbost...@gmail.com wrote: On Wed, 2010-11-24 at 12:51 -0500, Kingsley Idehen wrote: What does MySQL 4 do with this data that can't be done with a moderately capable RDF quad / triplestore? If I am going to run rings around this thing, I need a starting point :-) That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. This is about the 1000th time I have heard this story, and the ability range of those saying the same thing is huge - from 9-5 devs who learn what they need to people who research and teach artificial intelligence and other cutting edge areas and who actively learn new, complex skills just because they can. The point is not whether someone who (co?)developed the virtuoso triplestore can make RDF work, it's whether someone with the time, current documentation and inclination can. Ben -- Regards, Kingsley Idehen President CEO OpenLink Software Web: http://www.openlinksw.com Weblog: http://www.openlinksw.com/blog/~kidehen Twitter/Identi.ca:
CfP: Applied Ontology Special Issue on Modularity in Ontologies
= --- Applied Ontology --- --- Special Issue on Modularity in Ontologies --- --- Call for Papers --- Abstract Submission deadline: 31 December 2010 Paper Submission deadline: 31 January 2011 http://www.informatik.uni-bremen.de/~okutz/womoAO/ = Applied Ontology - An Interdisciplinary Journal of Ontological Analysis and Conceptual Modeling IOS Press, ISSN: 1570-5838 Editors-in-Chief: Nicola Guarino, Mark A. Musen MODULARITY, as studied for many years in software engineering, allows mechanisms for easy and flexible reuse, generalisation, structuring, maintenance, design patterns, and comprehension. Applied to ontology engineering, modularity is central not only to reduce the complexity of understanding ontologies, but also to facilitate ontology maintenance and ontology reasoning. Recent research on ontology modularity shows substantial progress in foundations of modularity, techniques of modularisation and modular development, distributed reasoning and empirical evaluation. These results provide a solid foundation and exciting prospects for further research and development. This special issue invites submission of high quality original work that has neither appeared in, nor is under consideration by, other journals. Submissions must be formatted according to IOS Press style ( www.iospress.nl/authco/instruction_crc.html ) and should be prepared in PDF format. Contributions must be received not later than December 2010 through the EasyChair Submission System ( www.easychair.org/account/signin.cgi?conf=aowomo11 ). IMPORTANT DATES Abstract Submission: till 31st December 2010 Paper Submission: closes 31st January 2011 Notification (expected): March/April 2011 Revised Version due: June 2011 Publication (expected): Fall 2011 GUEST EDITORS Oliver Kutz, Joana Hois (Research Center on Spatial Cognition (SFB/TR 8), Bremen, Germany) http://www.ontospace.uni-bremen.de
Re: Failed to port datastore to RDF, will go Mongo
On Wed, 24 Nov 2010 18:12:50 + Ben O'Steen bost...@gmail.com wrote: That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. Or, to put a different slant on it: a competent developer who has spent years using SQL databases day-to-day finds it easier to use SQL and the relational data model than a different data model and different query language that he's spent a few days trying out. It's not surprising. I often find it difficult to code things in Python and end up switching to Perl. Why? Is it because Perl's an inherently easier language? Or is it because Perl has been one of my main development tools for the best part of a decade whereas I dig out Python only occasionally. -- Toby A Inkster mailto:m...@tobyinkster.co.uk http://tobyinkster.co.uk
Re: Failed to port datastore to RDF, will go Mongo
* [2010-11-24 22:44:53 +] Toby Inkster t...@g5n.co.uk écrit: ] ] Or, to put a different slant on it: a competent developer who has spent ] years using SQL databases day-to-day finds it easier to use SQL and the ] relational data model than a different data model and different query ] language that he's spent a few days trying out. I don't think that's what's happening here, or at least not entirely. People coming from a RDB background expect things like SUM, COUNT, INSERT, DELETE, not to mention GROUP BY to work. But SPARQL 1.1 is still very new, each store implements them in slightly different ways with slightly different syntax, sometimes requiring workarounds in application code. With RDBs we have good libraries for abstracting away these differences. We still require people to pay a lot closer attention to what the underlying plumbing is and how it works (and if the binary package they got with their OS might be out of date or has to be compiled from source or even patched - the horror!). These things prevent people from getting on with what they see as the task at hand. Cheers, -w -- William Waites http://eris.okfn.org/ww/foaf#i 9C7E F636 52F6 1004 E40A E565 98E3 BBF3 8320 7664
Re: Failed to port datastore to RDF, will go Mongo
Hi all, first off: thanks a lot for the many comments and the advice that this has received. On Nov 24, 2010, at 11:44 PM, Toby Inkster wrote: On Wed, 24 Nov 2010 18:12:50 + Ben O'Steen bost...@gmail.com wrote: That's not the point that is being made. A competent developer, using all the available links and documentation, spending days researching and learning and trying to implement, is unable to make an app using a triplestore that is on a par with one they can create very quickly using a relational database. Or, to put a different slant on it: a competent developer who has spent years using SQL databases day-to-day finds it easier to use SQL and the relational data model than a different data model and different query language that he's spent a few days trying out. That is probably a fair description of my position. (Although I'm now using a database I've only known for 4 months and that is definitely not relational.) I want to finish this particular project by the end of next week, so I decided to default back to technologies that are more familiar to me and where it seemed easier to select the required components. Have a running prototype now :-) Anyway, I'd like to raise some additional points for the future: 1. I'd like to get a better picture of who is currently developing end-user open government data applications based on linked data. Given that there is a massive push towards releasing OGD as LD, I'd be eager to find out who is consuming it in which kind of (user-facing) context, especially regarding government transparency. More precisely: is RDF used primarily as an interchange format or are there many people actively running sites using it? 2. (Trying to figure out the intended process:) Several people have suggested that I shoud iteratively develop a mapping of the data to RDF, starting with an entirely independent ontology and then incrementally adopting other vocabularies. While this seems fine in theory, I'm curious how it works in practice: wouldn't I a) screw anyone using my data via REST and dumps and b) have to refactor all of my loaders, search indexing, templates, ... essentially every part of the system using the data for each change? Or would you recommend duplication? Thanks a lot, Friedrich