Re: Question on moving linked data sets

2012-04-19 Thread Ed Summers
Hi Antoine,

First, congratulations on http://data.bnf.fr/ that is a major milestone!

On Thu, Apr 19, 2012 at 10:23 AM, Antoine Isaac ais...@few.vu.nl wrote:
 We can ask for the people we know to change their links. But identifying the
 users of URIs seems too manual, error-prone a process. And of course in
 general we do not want links to be broken.

 Currently we have done the following:

 - a 301 moved permanently redirection from the stitch.cs.vu.nl/rameau
 prototype to data.bnf.fr.

 - an owl:sameAs statement between the prototype URIs and the production
 ones, so that a client searching for data on the old URI gets data that
 enables it to make the connection with the original resource (URI) it was
 seeking data about.

 Does that seem ok? What should we do, otherwise?

As you know when the SKOS concepts published at lcsh.info moved to
id.loc.gov I had a similar situation :-) Like you I chose to do a
mixture of technical and social things:

- publish information about the move to relevant discussion lists
- put some information up at lcsh.info about the move
- permanently redirect (301) all resources to their new location (easy
since it was the same app, and mod_rewrite could do it)
- after a year of redirects I shut down lcsh.info and did not renew
the domain (interestingly someone is squatting on it right now
attempting to sell it, I think)

Generally I think the linked data community should be encouraged to
check their links, respect 301 redirects, and update their own link
database appropriately. This is what Google and other major search
engines do [1], and it's how the Web was designed to work, and
continues to grow.

While it's certainly cool when URIs don't change [1] I think it is
somewhat irrational to expect URIs to be permanent. I hear people
gripe about broken URLs in the digital preservation community quite a
bit and it is pretty irritating, since any data that isn't actively
used tends to rot...URLs really aren't that different.

Of course there is certainly value in stable identifiers [1], but I
think there is an opportunity for documenting and encouraging best
practices on how to manage change (especially with respect to
identifiers) in Linked Data. URNs, Namespaces and Registries [2] is
partly helpful here, but a more succinct and URI focused presentation
is needed. Or maybe a best practice document like this already exists
and I haven't seen it yet. If that is the case I trust someone will
let me know :-)

//Ed

[1] http://www.w3.org/Provider/Style/URI.html
[2] http://www.w3.org/2001/tag/doc/URNsAndRegistries-50



Re: Question on moving linked data sets

2012-04-19 Thread Ed Summers
On Thu, Apr 19, 2012 at 11:02 AM, Bernard Vatant
bernard.vat...@mondeca.com wrote:
 My take on this would be to use dcterms:isReplacedBy links rather than
 owl:sameAs
 Description of the concepts by BNF might change in the future and although
 the original identifier is the same, the description might be out of sync at
 some point.

I really like the idea of avoiding the owl:sameAs quagmire with an
assertion that has less ontological consequences. When publishing data
about data.bnf resources perhaps it might be a bit simpler to use
dcterms:replaces instead, e.g.

http://data.bnf.fr/ark:/12148/cb14521343b dcterns:replaces
http://stitch.cs.vu.nl/vocabularies/rameau/ark:/12148/cb14521343b .

//Ed



Re: Question on moving linked data sets

2012-04-19 Thread Ed Summers
On Thu, Apr 19, 2012 at 1:23 PM, Ed Summers e...@pobox.com wrote:
 http://data.bnf.fr/ark:/12148/cb14521343b dcterns:replaces
 http://stitch.cs.vu.nl/vocabularies/rameau/ark:/12148/cb14521343b .

s/dcterns/dcterms/

//Ed



Re: PURLs don't matter, at least in the LOD world

2012-02-19 Thread Ed Summers
On Fri, Feb 17, 2012 at 7:02 PM, David Wood da...@3roundstones.com wrote:
 Given what I personally know of the state of US Government agencies, I'll 
 take your bet whether the Web services of the Library of Congress or OCLC 
 lasts longer :)  You might look back at the tortured history of id.loc.gov 
 before we agree to a figure.

At least w/ the tortured history of id.loc.gov and lcsh.info I was
able to permanently redirect lcsh.info to the appropriate places on
id.loc.gov when lcsh.info was slated for retirement. That way anybody
who was scrubbing their links (notably the search engines more than
the semantic web community) would have updated their links.

I'm with Hugh, putting all your identifier eggs in the basket of
purl.org (or any 3rd party service) isn't an excuse for not
thoughtfully managing your URL namespaces and DNS. Perhaps that's
tilting at windmills, but so be it. In my opinion more still needs to
be done to educate people about how web architecture actually works
instead of getting them to invest in niche software solutions,
maintained by a handful of people with consulting contracts on the
line.

//Ed



Re: Explaining the benefits of http-range14 (was Re: [HTTP-range-14] Hyperthing: Semantic Web URI Validator (303, 301, 302, 307 and hash URIs) )

2011-10-20 Thread Ed Summers
On Wed, Oct 19, 2011 at 12:59 PM, Leigh Dodds leigh.do...@talis.com wrote:
 So, can we turn things on their head a little. Instead of starting out
 from a position that we *must* have two different resources, can we
 instead highlight to people the *benefits* of having different
 identifiers? That makes it more of a best practice discussion and one
 based on trade-offs: e.g. this class of software won't be able to
 process your data correctly, or you'll be limited in how you can
 publish additional data or metadata in the future.

 I don't think I've seen anyone approach things from that perspective,
 but I can't help but think it'll be more compelling. And it also has
 the benefits of not telling people that they're right or wrong, but
 just illustrate what trade-offs they are making.

I agree Leigh. The argument that you can't deliver an entity like a
Galaxy to someone's browser sounds increasingly hollow to me. Nobody
really expects that, and the concept of a Representation from
WebArch/REST explains it away to most technical people. Plus, we now
have examples in the wild like OpenGraphProtocol that seem to be
delivering drinks, politicians, hotels, etc to machine agents at
Facebook just fine.

But there does seem to be a valid design pattern, or even refactoring
pattern, in httpRange-14 that is worth documenting. Perhaps a good
place would be http://patterns.dataincubator.org/book/? I think
positioning httpRange-14 as a MUST instead of a SHOULD or MAY made a
lot of sense to get the LOD experiment rolling. It got me personally
thinking about the issue of identity in a practical way as I built web
applications, that I probably wouldn't otherwise have otherwise done.
But it would've been easier if grappling with it was optional, and
there were practical examples of where it is useful, instead of having
it be an issue of dogma.

//Ed



REST and Linked Data (was Minting URIs)

2011-04-17 Thread Ed Summers
Hi Michael,

On Fri, Apr 15, 2011 at 1:57 PM, Michael Hausenblas
michael.hausenb...@deri.org wrote:
 I find [1] a very useful page from a pragmatic perspective. If you're more
 into books and not only focusing on the data side (see 'REST and Linked
 Data: a match made for domain driven development?' [2] for more details on
 data vs. API), I can also recommend [3], which offers some more practical
 guidance in terms of URI space management.

Thanks for the reference to the recent REST and Linked Data [1] paper,
I had missed it till now. I had some comments and questions, which I
figured couldn't hurt (too much?) to be discussed here:


Typically a REST service will assume the resource
being transferred in these representations can be considered
a document; in the terminology of the following section, an
‘information resource’.


I guess in some ways this is a minor quibble, but this seems to me to
be a fairly common misconception of REST in the Linked Data community.
In his dissertation Roy says this about resources [2]:


The key abstraction of information in REST is a resource. Any
information that can be named can be a resource: a document or image,
a temporal service (e.g. today's weather in Los Angeles), a
collection of other resources, a non-virtual object (e.g. a person),
and so on.


It is actually pretty common to use URLs to identify real world
things when designing a REST API for a domain model. However, if a
developer implements some server side functionality to allow (for
example) a product to be deleted with an HTTP DELETE to the right URL,
they would typically be more concerned with practical issues such as
authentication, than with the existential ramifications of deleting
the product (should all instances of the product be recalled and
destroyed?).

Section 3.3 of the REST and Linked Data paper goes on to say:


Linked Data services, in implementing the “HTTP range
issue 14” solution [12], add semantics to the content negoti-
ation to distinguish between URIs that are non-information
resources (identifiers for conceptual or real-world objects)
and URIs that are information resources (documents) that
describe the non-information resources. This is because as-
sertions in the RDF graph are usually relationships that ap-
ply to the non-information resource, but Linked Data over-
loads URI usage so that it is also a mechanism for retrieving
triples describing that resource (in a document, i.e. an in-
formation resource). (This is a change iin behaviour from
earlier use of HTTP URIs in RDF, when they were not ex-
pected to be dereferenced.)


Was there really a chapter in the history of the Semantic Web where
HTTP URIs in RDF were not expected to resolve? Wouldn't that have
removed all the machinery for the Web from the Semantic Web? I
only really started paying attention to RDF when the Linked Data
effort (SWEO) picked up, so I'm just generally interested in this
pre-history, and other people's memory of it.

Minor quibbles aside, as a web developer I'm a big supporter of the
paper's conclusions, which promote REST's notion of Hypertext As The
Engine Of Application State (HATEOS), and the semantics that the HTTP
Connector provides, in the Linked Data space:


3. Services that publish Linked Data resources should pay
careful consideration to HATEOAS as a viable altern-
ative SPARQL, and identify resources to enable REST-
ful use of the API.
4. RESTful methods should be developed for the write-
enabled Web of Data.


Thanks to the authors (cc'd here) for writing that down and presenting
it forcefully.

//Ed

[1] http://ws-rest.org/2011/proc/a5-page.pdf
[2] 
http://www.ics.uci.edu/~fielding/pubs/dissertation/rest_arch_style.htm#sec_5_2



Re: [ANN] Linked Open Colors

2011-04-01 Thread Ed Summers
Hi Sergio,

Thanks for the update about this new, much needed service! While
perusing the data from the command line with rapper I noticed that:

http://data.colourphon.co.uk/id/colourscheme/triadic/ff6347
rdfs:label triadic colours to compliment Coral (ff6347) .

ISO 01.070 [1] clearly defines this color as Tomato. I think we can
all agree that the identity of colors is a solved problem, and it
would be great if this Linked Data service could be updated to reflect
this. I started a Doodle Poll to see if we can find a time to meet to
discuss whether a W3C Working Group or Incubator Group might be needed
to resolve this issue.

//Ed

[1] http://bit.ly/gWCSql
[2] http://www.doodle.com/geyvyrx4wzgvky5s



Re: Low Quality Data (was before Re: AW: ANN: LOD Cloud - Statistics and compliance with best practices)

2010-10-26 Thread Ed Summers
Somewhat related: the CPAN folks coined their own term for code
quality (kwalitee), with precise measures, which takes some of the
subjectiveness out of the equation:

http://qa.perl.org/phalanx/kwalitee.html

Perhaps something like this could work for linked data quality?

//Ed



Re: empirical cloud

2010-09-21 Thread Ed Summers
Hi Antoine:

On Tue, Sep 21, 2010 at 3:58 AM, Antoine Isaac ais...@few.vu.nl wrote:
 By complete chance I spotted a small bibsonomy.org node attached to
 semanticweb.org through dblp.l3s.de. But there is also a www.bibsonomy.org
 node (much bigger). Are they supposed to be the same? Are there other cases?

No they're not supposed to be the same, at least in the little
visualization I did. If there's a sensible heuristic for merging them
I'd be willing to try it out though.

//Ed



empirical cloud

2010-09-15 Thread Ed Summers
The discussion about updating the Linking Open Data cloud got me
thinking it would be interesting to try to visualize the actual
owl:sameAs links in the Billion Triple Challenge dataset [1]. It
turned out to be relatively easy to get something somewhat workable
with tools like zgrep, sort, uniq, a couple of custom scripts, and the
handy ProtoVis library [2].

You can view the result at:

  http://inkdroid.org/empirical-cloud/

If you notice the links to the rdf/xml and Turtle you'll see I tried
my hand at representing the underlying data using void, foaf and
dcterms. There's definitely room for improvement, so if you are so
inclined please tinker with it and send me a pull request on GitHub
[3]. Thanks to Gunnar Grimnes and Dan Brickley for ideas and help.

//Ed

[1] http://challenge.semanticweb.org/
[2] http://vis.stanford.edu/protovis/
[3] http://github.com/edsu/empirical-cloud



Re: LOB - linked open bard

2010-09-04 Thread Ed Summers
Hi Toby,

Do you have idea where the file URIs are coming from?

  rapper -o turtle http://ontologi.es/lob/

//Ed

On Fri, Sep 3, 2010 at 7:44 PM, Toby Inkster t...@g5n.co.uk wrote:
 Every play by Shakespeare in RDF:

        http://ontologi.es/lob/

 For example:

        http://ontologi.es/lob/Hamlet.ttl

 Two Noble Kinsmen is not included, as I couldn't find a good XML
 version to convert to RDF, and Shakespeare only co-wrote it anyway.



Re: Semantic Web at SXSW

2010-07-03 Thread Ed Summers
I wonder if it might be worthwhile approaching Manu Sporny about
putting in a proposal to talk about rdfa, microformats, his recent
work with json, or something else relevant/timely that he thinks might
fit?

As Dan Brickley pointed out recently [1], the semweb/ linkeddata crowd
really share 99.9% of their geek DNA with the microformats
community...and Manu has done a really great job bridging the two
communities. Microformats have been a popular topic at past sxsw
conferences.

Thanks for continuing to push on this Juan.

//Ed

[1] http://lists.w3.org/Archives/Public/public-lod/2010Jul/0020.html



Fwd: Free Linked Data Webinar

2010-05-23 Thread Ed Summers
This could be of potential interest to Linked Data folks who work in
cultural heritage institutions (libraries, archives, musuems, etc).

//Ed

-- Forwarded message --
From: Roy Tennant roytenn...@gmail.com
Date: Fri, May 21, 2010 at 5:15 PM
Subject: [CODE4LIB] Free Linked Data Webinar
To: code4...@listserv.nd.edu

My colleague Ralph LeVan, whom many of you know, is giving a free webinar on
Linked Data in our OCLC Research webinar series Technical Advances for
Innovation in Cultural Heritage Institutions (TAI CHI). It will be on 27
May at 2PM EDT. To find out more and to register go here:

http://www.oclc.org/research/news/2010-05-21.htm

Also, there will be a recording made available afterwards. Anyone who has
heard Ralph speak knows that he is gifted at making complex topics
understandable. So if you're wondering about what all the fuss is about
regarding Linked Data, this will be something not to miss.

Thanks,
Roy Tennant
OCLC Research



Re: rdf/json

2010-05-03 Thread Ed Summers
Perhaps this has already been suggested, but the linked-data-api folks
have an overview of json serializations for rdf [1] that might be of
some interest.

//Ed

[1] http://code.google.com/p/linked-data-api/wiki/JSONFormats



Re: Linked data in packaged content (ePub)

2010-04-28 Thread Ed Summers
On Tue, Apr 27, 2010 at 4:40 PM, Stuart A. Yeates syea...@gmail.com wrote:
 Does anyone know of any other attempts to put linked data into
 packages like this?

While arguably not Linked Data per-se, you might be interested in work
being done on the Open Publication Distribution System (OPDS) [1],
which aims to use Atom for making metadata about ebooks available. A
key part of an OPDS feed are opds:acquisition links between an
atom:entry and a epub document identified with a URI and a media type,
for example:

link type=application/epub+zip
href=http://www.feedbooks.com/book/4440.epub;
rel=http://opds-spec.org/acquisition/

Implementors include people at O'Reilly, Internet Archive, Ibis
Reader, FeedBooks, to name a few. Much of the work is actually going
on in open bi-weekly conference calls, and on a discussion list [3] if
you are interested. On a recent call Hadrien Gardeur of Feedbooks was
talking about embedding opds atom documents in the ebook
serializations, so it might be worthwhile pinging him and/or the
discussion list to see where things are at.

//Ed

[1] http://code.google.com/p/openpub/wiki/CatalogSpecDraft
[2] http://ibisreader.com/
[3] http://groups.google.com/group/openpub



Re: LOD Camp at WWW2010....

2010-04-20 Thread Ed Summers
Hi Ivan,

So it looks like students could go to the entire day of Linked Open
Data Camp by paying $75.00?

//Ed

On Tue, Apr 20, 2010 at 12:19 PM, Ivan Herman i...@w3.org wrote:
 This is just a reminder that there will be an LOD camp at the WWW Conference 
 next week in Raleigh:

 http://www.w3.org/2010/04/w3c-track.html

 you should sign up on

 http://esw.w3.org/topic/LODCampW3CTrack

 you can also add your favourite topic right now on the wiki page, or leave 
 for the discussion over there...

 Ivan

 
 Ivan Herman, W3C Semantic Web Activity Lead
 Home: http://www.w3.org/People/Ivan/
 mobile: +31-641044153
 PGP Key: http://www.ivan-herman.net/pgpkey.html
 FOAF: http://www.ivan-herman.net/foaf.rdf









Re: Hungarian National Library published its entire OPAC and Digital Library as Linked Data

2010-04-13 Thread Ed Summers
Sorry I had a small typo in that oai-ore example I included. I meant
to type ore:aggregates instead of ore-aggregates. I also meant to
include assertions about the format of the files, which can be handy
to have. //Ed

http://oszkdk.oszk.hu/resource/DRJ/404
   dc:creator http://nektar.oszk.hu/resource/auth/33589, Jókai
Mór,, (1825-1904.) ;
   dc:date cop. 2006 ;
   dc:description 
Működési követelmények: Adobe Reader / MS Reader
, Főcím a címképernyőről, Szöveg (pdf : 1.2 MB) (lit : 546 KB) ;
   dc:identifier , 963-606-169-6 (pdf), 963-606-170-X (lit) ;
   dc:language hun ;
   dc:publisher Szentendre : Mercator Stúdió ;
   dc:subject http://nektar.oszk.hu/resource/auth/magyar_irodalom,
magyar irodalom. ;
   dc:title Dekameron ;
   dc:type book, elbeszélés., elektronikus dokumentum., no
type provided ;
   ore:aggregate
http://oszkdk.oszk.hu/storage/00/00/04/04/dd/1/dekameron.pdf,
http://oszkdk.oszk.hu/storage/00/00/04/04/dd/2/dekameron.lit .

http://oszkdk.oszk.hu/storage/00/00/04/04/dd/1/dekameron.pdf
dc:format application/pdf .
http://oszkdk.oszk.hu/storage/00/00/04/04/dd/2/dekameron.lit
dc:format application/x-ms-reader .



Re: CoIN: Composition of Identifier Names

2010-04-13 Thread Ed Summers
2010/4/13 Richard Cyganiak rich...@cyganiak.de:
 I think that URI Templates [3] might be a handy companion syntax for CoIN
 and I wonder if they could be integrated into CoIN. I'm thinking more about
 the general curly-brace-syntax rather than the fancy details. So perhaps you
 could start with something like

 http://example.org/publ/{publisher}/{document}
 http://example.org/publ/{publisher}/{document}/rev/{date}
 http://example.org/profiles/{name}

I second the idea of exploring the use of URI Templates for
documenting how to construct a URL from other data. I'm not sure if
it's part of the latest URI Templates draft [1], but OpenSearch allows
parameter names to be defined with namespaces [2]. For example:

 ?xml version=1.0 encoding=UTF-8?
 OpenSearchDescription xmlns=http://a9.com/-/spec/opensearch/1.1/;

xmlns:geo=http://a9.com/-/opensearch/extensions/geo/1.0/;
   Url type=application/vnd.google-earth.kml+xml
   
template=http://example.com/?q={searchTerms}pw={startPage?}bbox={geo:box?}format=kml/
/OpenSearchDescription

Note, the use of the geo namespace and the geo:box parameter name? So
you could imagine a URL template that referenced names from an RDF
vocabulary:

Url type=application/rdf+xml
template=http://example.com/user/{foaf:mbox_sha1sum}; /

OpenSearch was an incubator for the ideas that led to the URI
Templates draft, and is built into many modern web browsers (IE,
Firefox, Chrome).

//Ed

[1] http://tools.ietf.org/html/draft-gregorio-uritemplate-04
[2] http://www.opensearch.org/Specifications/OpenSearch/1.1#Parameter_names



Re: Using predicates which have no ontology?

2010-04-05 Thread Ed Summers
Hi Michael,

Would it be hard to remove the empty literal assertions? e.g.

--
http://www.iana.org/assignments/relation/alternate a awol:RelationType ;
 rdfs:label alternate ;
 dcterms:dateAccepted  ;
 dcterms:description  ;
 rdfs:isDefinedBy http://www.iana.org/go/rfc4287 .
--

It's interesting that the latest efforts to create a Link Relation
Registry seem to be intentionally avoiding publishing machine readable
data for the registry [1]. I was wondering if Mark Nottingham's
efforts to revamp link relations might present a good opportunity for
us to lobby the IETF to start publishing a bit of RDFa for the link
relations registry...

//Ed

[1] http://tools.ietf.org/html/draft-nottingham-http-link-header-09#appendix-A



JISC Grant Funding for Linked Data

2010-03-09 Thread Ed Summers
I just noticed that a recent JISC Grant Funding Call [1] includes some
funding available for Linked Data publishing projects by higher
education institutions in the UK:


Strand B - Expose

Projects that enable content to be made available on the Web using
structured data, in particular linked data which increases its
potential value to researchers, teachers and learners.

Total funds: £750,000. Up to 20 projects will be funded. Maximum
funding for any one project is £100,000.

A Briefing Document is available alongside this call, which outlines
key background and scope information to which bidders should refer.

The deadline for receipt of proposals in response to this call is 12
noon UK time on Tuesday 20 April 2010.


//Ed

[1] 
http://www.jisc.ac.uk/fundingopportunities/funding_calls/2010/03/210depositexpose.aspx



Re: Linking HTML pages and data

2010-02-16 Thread Ed Summers
Hi Sean,

On Tue, Feb 16, 2010 at 7:39 AM, Sean Bechhofer
sean.bechho...@manchester.ac.uk wrote:
 A simple (possibly dumb) question. Is there a standard mechanism for linking
 an HTML page to the non-information resource that it describes?

Not a dumb question at all--at least for me :-)

I've been using the link pattern that Chris Bizer, Richard Cyganiak
and Tom Heath documented in How To Publish Linked Data On The Web [1]
for discovery of RDF documents that are related to a given HTML
document.

But you are asking about the relation between a document and the the
*thing* being described. I agree with Damien that foaf:primaryTopic
seems like it could work, and that one possibility would be to slip a
bit of RDFa into the HTML document that asserted:

   foaf:primaryTopic http://dbpedia.org/resource/Mogwai_(band) .

I also agree w/ Kingsley that it would be neat to also have a link
pattern that non-RDFa folks could use:

  link rel=http://xmlns.com/foaf/0.1/primaryTopic;
href=http://dbpedia.org/resource/Mogwai_(band) title=Mogwai /

Or if mnot's Web Linking RFC is approved it would open the door to
using the Link HTTP Header:

  Link: http://dbpedia.org/resource/Mogwai_(band);
rel=http://xmlns.com/foaf/0.1/primaryTopic;; title=Mogwai

Registering [3] primaryTopic as a link relation type would tighten it
up a bit, as well as help advertise the pattern.

  link rel=primaryTopic
href=http://dbpedia.org/resource/Mogwai_(band) title=Mogwai /

At any rate, I'd be interested to hear if other people have other
approaches to this. It would be nice to have a little recipe (w3c
note?) people could follow when making these sorts of assertions on
the web. Assuming one isn't there already of course :-)

//Ed

PS. the oai-ore folks had a similar use case to link descriptions to
the thing being described. They ended up creating a new term
oai:describes [4], and documented ways of layering assertions into rdf
[5], atom [6] and html [7] documents. I think the vocabulary is
probably too specific to aggregations and resource maps to be useful
in the general case you are talking about though.

PSS. I really just wanted to type Mogwai a bunch of times :-)

[1] http://www4.wiwiss.fu-berlin.de/bizer/pub/LinkedDataTutorial/#discovery
[2] http://tools.ietf.org/html/draft-nottingham-http-link-header-07
[3] http://tools.ietf.org/html/draft-nottingham-http-link-header-07#section-6.2
[4] http://www.openarchives.org/ore/1.0/vocabulary#ore-describes
[5] http://www.openarchives.org/ore/1.0/rdfxml#remtoaggr
[6] http://www.openarchives.org/ore/1.0/atom#metadata
[7] http://www.openarchives.org/ore/1.0/discovery#HTMLLinkElement



Re: Linking HTML pages and data

2010-02-16 Thread Ed Summers
On Tue, Feb 16, 2010 at 5:51 PM, Ian Davis li...@iandavis.com wrote:
 You can see it in use on data.gov.uk:

 http://education.data.gov.uk/doc/school/56

 contains:

 link rel=primarytopic href=http://education.data.gov.uk/id/school/56; 
 /

Wow, thanks Ian. I hadn't noticed this pattern in use at data.gov.uk.
It seems like a worthwhile pattern to encourage people to follow, by
adding it to the How to Publish Linked Data on the Web [1] ... or
elsewhere?

//Ed

[1] http://www4.wiwiss.fu-berlin.de/bizer/pub/LinkedDataTutorial/



Re: [pedantic-web] ANN: 20th Century Press Archives as ORE / Linked Data application - Technical Preview

2009-12-28 Thread Ed Summers
On Mon, Dec 28, 2009 at 8:07 AM, Neubert Joachim j.neub...@zbw.eu wrote:
 Please feel invited to take a look at it - we would highly appreciate any
 feedback about our approach.

Thanks for announcing this Joachim. It is great to see more linked
data as rdfa getting out on the web. I'm particularly excited because
of your use of the oai-ore vocabulary to make historic newspaper
archives available, since we are doing something similar at the
Library of Congress [1].

You must've done something right because I just wrote a little naive
crawler [2] in a matter of minutes to pull down what looks like all
the rdfa you've put out there so far. It seem to have collected about
11,427 triples [3]. My rdfsum unix command line hack [4] came up with
these rdf:type counts:

   1533 http://www.openarchives.org/ore/terms/AggregatedResource
526 http://www.openarchives.org/ore/terms/ResourceMap
526 http://www.openarchives.org/ore/terms/Aggregation
336 http://zbw.eu/namespaces/skos-extensions/PmPage
185 http://purl.org/ontology/bibo/Article
  2 http://zbw.eu/namespaces/skos-extensions/PmPersonFolder
  2 http://zbw.eu/namespaces/skos-extensions/PmCollection

Does that sound about right for this initial release?

I noticed that you have chosen to link to names in the German National
Authority file like:

  http://zbw.eu/beta/pm20/person/00012 dct:subject
http://d-nb.info/gnd/118646419 .

I seem to remember hearing at SWIB09 [5] that the Deutsche National
Bibliothek was thinking about minting URIs for entries in the
authority file that follow Linked Data best practices (hash or 303,
etc). Were you planning on modifying these appropriately when those
URLs became available? Right now the d-nb URL returns 200 OK, and it
isn't a hash URI. Theoretically it would be pretty easy to layer in
some rdfa into the page at d-nb that describes:

  http://d-nb.info/gnd/118646419#person

But I realize this is somewhat out of your control. I guess it would
also be possible to create a partial PURL [6] for
http://d-nb.info/gnd/ that would redirect, since I think the new PURL
software supports 303.

I was also interested to see that you have published some SKOS
Extensions [7] that are used to type each ore:Aggregation as a
specialization of skos:Concept:

http://zbw.eu/beta/pm20/person/00012
a ore:Aggregation,
http://zbw.eu/namespaces/skos-extensions/PmPersonFolder ;
skos:prefLabel Abbe, Ernst; 1840-1905 (PM20 Personenarchiv)@de,
Abbe, Ernst; 1840-1905 (PM20 Persons Archives)@en .

It looks like the rdf that comes back for your skos extensions
vocabulary (nice hack with the rdf validator btw) doesn't define
PmPersonFolder--but perhaps I missed it? I'm guessing from the
skos:prefLabel assertion that the PmPersonFolder is a specialization
of skos:Concept?

Would it be OK for me to experiment with pulling down the aggregated
resource bitstreams (jpg, etc) and storing them on disk? It would just
be a single threaded little script. Part of the rationale behind the
ore use at LC [1] is to foster LOCKSS [8] scenarios where digital
objects are easier to meaningfully harvest.

Anyhow, I have rattled on enough for now I suppose -- I mainly wanted
to say how exciting it was to see your announcement, being from the
digital library tribe in the linked data community :-)

//Ed

[1] http://chroniclingamerica.loc.gov
[2] http://inkdroid.org/bzr/ptolemy/crawl.py
[3] http://inkdroid.org/data/pm20.txt
[4] http://inkdroid.org/bzr/bin/rdfsum
[5] http://www.swib09.de/
[6] http://purl.org
[7] http://zbw.eu/namespaces/skos-extensions/
[8] http://en.wikipedia.org/wiki/LOCKSS



Re: RDF Update Feeds

2009-11-20 Thread Ed Summers
At the Library of Congress we've been experimenting with using an Atom
feed to alert subscribers to new resources available at id.loc.gov
[1]. The approach is similar to what Niklas' is doing, although we
kind of independently arrived at this approach (which was nice to
discover).

Creates, updates and deletes happen on a weekly basis, so it's
important for us to let interested parties know what has changed. We
ended up using Atom Tombstones [2] for representing the deletes. And
Atom Feed Paging and Archiving (RFC 5005) [3]  to allow clients to
drill backwards through time.

I just noticed Link Relations for Simple Version Navigation [4] get
announced on an Atom related discussion list, which looks like it
could be useful as well, if you maintain a version history.

I'd be interested in any feedback anyone has about using this approach.

//Ed

[1] http://id.loc.gov/authorities/feed/
[2] http://ietfreport.isoc.org/all-ids/draft-snell-atompub-tombstones-06.txt
[3] http://tools.ietf.org/rfc/rfc5005.txt
[4] http://www.ietf.org/id/draft-brown-versioning-link-relations-03.txt



Re: RDF Update Feeds

2009-11-20 Thread Ed Summers
On Fri, Nov 20, 2009 at 11:05 AM, Nathan nat...@webr3.org wrote:
 is this not the same as (or vi similar to) the court approach outlined
 here: http://code.google.com/p/court/ by Niklas

Yes, absolutely. Although I had no idea of Niklas' work at the time.
That's why I said:


At the Library of Congress we've been experimenting with using an Atom
feed to alert subscribers to new resources available at id.loc.gov
[1]. The approach is similar to what Niklas' is doing, although we
kind of independently arrived at this approach (which was nice to
discover).


:-)

//Ed



Re: Top three levels of Dewey Decimal Classification published as linked data

2009-08-20 Thread Ed Summers
On Thu, Aug 20, 2009 at 3:53 AM, Bernard
Vatantbernard.vat...@mondeca.com wrote:
 http://dewey.info/class/641/
    cc:attributionName OCLC Online Computer Library Center, Inc. ;
    cc:attributionURL http://www.oclc.org/dewey/ ;
    cc:morePermissions http://www.oclc.org/dewey/about/licensing/ ;
    dct:hasVersion http://dewey.info/class/641/2009/08/ ;
    dct:language de^^dct:RFC4646 ;
    a skos:Concept ;
    xhtml:license http://creativecommons.org/licenses/by-nc-nd/3.0/ ;
    skos:broader http://dewey.info/class/64/2003/08/about.de ;
    skos:inScheme http://dewey.info/scheme/2003/08/about.de ;
    skos:notation 641^^http://dewey.info/schema-terms/Notation ;
    skos:prefLabel Food  drink@en, Essen und Trinken@de .


 Maybe rather :

 skos:broader http://dewey.info/class/64/2003/ ;
 skos:inScheme http://dewey.info/scheme/2003/08/ ;

Yes, thanks Bernard. In my haste I forgot to remove the language
specific-ness from the skos:broader assertions in my email to Michael.
I might actually recommend that the version information is removed
from the object in the broader relation too:

http://dewey.info/class/641/ skos:broader
http://dewey.info/class/64/2003/ .

As it stands now a document URI is being used in the object position,
instead of the URI for the concept.

uqbar:~ ed$ curl -I http://dewey.info/class/64/2003/08/about.de
HTTP/1.1 200 OK
Date: Thu, 20 Aug 2009 09:27:32 GMT
Server: Apache/2.2.11 (Unix) PHP/5.2.10
X-Powered-By: PHP/5.2.10
Content-Location: http://dewey.info/class/64/2003/08/about.de.rdf
Vary: Accept
Content-Length: 2115
Content-Type: application/rdf+xml

//Ed



Re: Top three levels of Dewey Decimal Classification published as linked data

2009-08-19 Thread Ed Summers
Hi Michael,

This is really exciting news, especially for those of us Linked Data
enthusiasts in the world of libraries and archives. Congratulations!

I haven't fully read the wiki page yet, so I apologize if this
question is already answered there. I was wondering why you chose to
mint multiple URIs for the same concept in different languages. For
example:

uqbar:~ ed$ rapper -o turtle http://dewey.info/class/641/

@prefix rdf: http://www.w3.org/1999/02/22-rdf-syntax-ns# .
@prefix xhtml: http://www.w3.org/1999/xhtml/vocab# .
@prefix cc: http://creativecommons.org/ns# .
@prefix dct: http://purl.org/dc/terms/ .
@prefix skos: http://www.w3.org/2004/02/skos/core# .

http://dewey.info/class/641/2009/08/about.en
cc:attributionName OCLC Online Computer Library Center, Inc. ;
cc:attributionURL http://www.oclc.org/dewey/ ;
cc:morePermissions http://www.oclc.org/dewey/about/licensing/ ;
dct:isVersionOf http://dewey.info/class/641/ ;
dct:language en^^dct:RFC4646 ;
a skos:Concept ;
xhtml:license http://creativecommons.org/licenses/by-nc-nd/3.0/ ;
skos:broader http://dewey.info/class/64/2009/08/about.en ;
skos:inScheme http://dewey.info/scheme/2009/08/about.en ;
skos:notation 641^^http://dewey.info/schema-terms/Notation ;

http://dewey.info/class/641/2003/08/about.de
cc:attributionName OCLC Online Computer Library Center, Inc. ;
cc:attributionURL http://www.oclc.org/dewey/ ;
cc:morePermissions http://www.oclc.org/dewey/about/licensing/ ;
dct:isVersionOf http://dewey.info/class/641/ ;
dct:language de^^dct:RFC4646 ;
a skos:Concept ;
xhtml:license http://creativecommons.org/licenses/by-nc-nd/3.0/ ;
skos:broader http://dewey.info/class/64/2003/08/about.de ;
skos:inScheme http://dewey.info/scheme/2003/08/about.de ;
skos:notation 641^^http://dewey.info/schema-terms/Notation ;
skos:prefLabel Essen und Trinken@de .

...

I kind of expected the assertions to hang off of a language and
version agnostic URI, with perhaps dct:hasVersion links to previous
versions.

http://dewey.info/class/641/
cc:attributionName OCLC Online Computer Library Center, Inc. ;
cc:attributionURL http://www.oclc.org/dewey/ ;
cc:morePermissions http://www.oclc.org/dewey/about/licensing/ ;
dct:hasVersion http://dewey.info/class/641/2009/08/ ;
dct:language de^^dct:RFC4646 ;
a skos:Concept ;
xhtml:license http://creativecommons.org/licenses/by-nc-nd/3.0/ ;
skos:broader http://dewey.info/class/64/2003/08/about.de ;
skos:inScheme http://dewey.info/scheme/2003/08/about.de ;
skos:notation 641^^http://dewey.info/schema-terms/Notation ;
skos:prefLabel Food  drink@en, Essen und Trinken@de .

See how multiple skos:prefLabel assertions can be made using the same
subject? To illustrate why I think this is important consider how
someone may use the resources at dewey.info:

http://openlibrary.org/b/OL11604988M dct:subject
http://dewey.info/class/641/ .

If we follow our nose to http://dewey.info/class/641 we will get back
some RDF, but the description we get isn't about the subject
http://dewey.info/class/641/ so what are we to make of the above
assertion?

Another unrelated thing I noticed is that the RDFa doesn't seem to be
usable by the RDFa Distiller:

http://tinyurl.com/nm8lfa

Sorry if this was too long. Let me just say again how exciting it is
to see this work coming out of OCLC.

//Ed



Chronicling America and Linked Data pt. 2

2009-05-26 Thread Ed Summers
I forgot to mention that if you are interested in reading more about
the use of the Object Reuse and Exchange vocabulary (OAI-ORE) in
linked data, and web based digital repositories you'll want to check
out this recent paper from LDOW2009:

  H. Van de Sompel, C. Lagoze, M. Nelson, S. Warner, R. Sanderson, P.
Johnston. Adding eScience Assets to the Data Web. Linked Data on the
Web (LDOW2009). April 20th 2009. Madrid, Spain.
http://events.linkeddata.org/ldow2009/papers/ldow2009_paper8.pdf

//Ed



Chronicling America and Linked Data

2009-05-26 Thread Ed Summers
There is a new pool of linked-data up at the Library of Congress in
the Chronicling America application [1]. Chronicling America is the
web view on data collected for the National Digital Newspaper Program
(NDNP). NDNP is a 20-year joint project of the National Endowment for
the Humanities and the Library of Congress to digitize and aggregate
historic newspaper in the United States.

Right now there are close to a million digitized newspaper pages
available, and information about 140,000 newspaper titles...all of
which have individual web views, for example:

 Newspaper Title: San Francisco Call [2]
 Issue: San Francisco Call, 1895-03-05 [3]
 Page: San Francisco Call, 1895-03-05, page sequence 1 [4]

If you view source on them you should be able to auto-discover the
application/rdf+xml representations that bundle up information about
the newspaper titles, issues and pages. You can also browse around
using a linked data viewer like uriburner [5].

The implementation is a moving target, but you'll see we've cherry
picked a few vocabularies: Dublin Core [6], Bibliographic Ontology
[7], FOAF [8], and Object Reuse and Exchange (OAI-ORE) [9]. ORE in
particular was extremely useful to us, since we wanted to enable the
application's repository function, by exposing the digital objects
(image files, ocr/xml files, pdfs) that make up the individual Page
resources. For example:

http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1#page
ore:aggregates
http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1.jp2,
http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1.pdf,
http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1/ocr.txt,
http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1/ocr.xml,
http://chroniclingamerica.loc.gov/lccn/sn84026749/1905-01-29/ed-1/seq-1/thumbnail.jpg
.

The idea is to enable the harvesting of these repository objects out
of the Chronicling American webapp. The only links out we have so far
are from Newspaper Titles to the geographic regions that they are
about, and languages. So for example:

http://chroniclingamerica.loc.gov/lccn/sn85066387#title
dcterms:coverage
http://dbpedia.org/resource/San_Francisco%2C_California,
http://sws.geonames.org/5391959/ ;
dcterms:language http://www.lingvoj.org/lang/en .

Just these minimal links provide a huge amount of data enrichment to
our original data. We also needed to create a handful of new
vocabulary terms, which we made available as RDFa [10]. I would be
interested in any feedback you have. Also, please feel free to fire up
linked-data bots to crawl the space.

//Ed

[1] http://chroniclingamerica.loc.gov
[2] http://chroniclingamerica.loc.gov/lccn/sn85066387/
[3] http://chroniclingamerica.loc.gov/lccn/sn85066387/1895-03-05/ed-1/
[4] http://chroniclingamerica.loc.gov/lccn/sn85066387/1895-03-05/ed-1/seq-1/
[5] 
http://linkeddata.uriburner.com/about/html/http/chroniclingamerica.loc.gov/lccn/sn84026749%23title
[6] http://dublincore.org/
[7] http://bibliontology.com/
[8] http://xmlns.com/foaf/spec/
[9] http://www.openarchives.org/ore/1.0/vocabulary.html
[10] http://chroniclingamerica.loc.gov/terms/



Re: Chronicling America and Linked Data pt. 2

2009-05-26 Thread Ed Summers
On Tue, May 26, 2009 at 12:33 PM, Dan Brickley dan...@danbri.org wrote:
 BTW Ed, Chronicling America looks great. Nice work as ever :) Is there any
 people-describing data in there, or will there be?

Thanks Dan. It's funny you ask, because so far the predominant users
of Chronicling America have been genealogists looking for people :-)

At the moment (apart from metadata about newspaper titles, and issues)
we only have OCR for the pages, which allows us to see what words
appear where on the page for indexing and highlighting. We definitely
want to do more mining of the text, looking for names, dates, places,
etc embedded in the articles. Part of the hope is that by releasing
the data interested people can mine it themselves, and publish
assertions like:

  http://chroniclingamerica.loc.gov/lccn/sn86069133/1909-02-02/ed-1/seq-7#page
dct:subject http://dbpedia.org/resource/Abraham_Lincoln .

Perhaps we'll eventually get around to providing functionality in the
app that makes it easy to do this sort of annotation. At the moment
there aren't any plans to do so, but we're already tossing around
ideas for a new user interface, so I am hopeful.

If you have any ideas please feel free to send them here or to me privately.

//Ed



Re: Suggestions/existing material for Linked Data tutorial?

2008-11-14 Thread Ed Summers

On Fri, Nov 14, 2008 at 6:40 AM, Yves Raimond [EMAIL PROTECTED] wrote:
 We plan to publish, with Keith, a small how-to for doing a hands-on
 tutorial like the Web-of-data 101 session we did at the WOD-PD event.
 With Richard's permission, we'll also take a few things from his
 session, where he used some of the data we created during our session.

 The goal of this how-to is to make it easy for people to reproduce
 the tutorial in different places.

This would be extremely helpful to me and some others who are thinking
of doing a linked-data tutorial at code4lib2009 [1]. Please post a
message here when any of the materials materialize!

//Ed

[1] http://code4lib.org/2009



Re: Linked data at Freebase

2008-10-29 Thread Ed Summers

Hi Yves:

Wow, this is exciting news! I wonder would it be a quick fix to get
the service doing a 303 instead of a 302:

uqbar:~ ed$ curl -I http://rdf.freebase.com/ns/en.blade_runner
HTTP/1.0 302 FOUND
Server: PasteWSGIServer/0.5 Python/2.4.4
Date: Wed, 29 Oct 2008 04:10:11 GMT
content-type: text/html; charset=utf-8
location: http://rdf.freebase.com/rdf/en/blade_runner
pragma: no-cache
cache-control: no-cache
Connection: close

Also, could the HTML views (e.g.
http://www.freebase.com/view/en/blade_runner) have a bit of
autodiscovery for the data view?

head
...
link rel=alternate type=application/turtle
href=http://rdf.freebase.com/rdf/en/blade_runner; title=rdf /
   ...
/head

This would allow agents that care about the RDF view to be able to
discover it while browsing the web of documents.

Nice work!

//Ed

PS. You can use Vapour the Linked Data Validator to help with some of
this checking: for example
http://idi.fundacionctic.org/vapour?vocabUri=http%3A%2F%2Frdf.freebase.com%2Fns%2Fen.blade_runnerclassUri=http%3A%2F%2FpropertyUri=http%3A%2F%2FinstanceUri=http%3A%2F%2FdefaultResponse=dontmind

On Wed, Oct 29, 2008 at 4:37 AM, Yves Raimond [EMAIL PROTECTED] wrote:

 Hello!

 Live from ISWC2008, I thought most of the people on this list would be
 interested in that:

 http://rdf.freebase.com/

 Freebase now produces linked data! (now, to see whether it links to
 other datasets :-) )

 Cheers!
 y





Re: Drilling into the LOD Cloud

2008-09-22 Thread Ed Summers

On Mon, Sep 22, 2008 at 1:28 PM, Giovanni Tummarello
[EMAIL PROTECTED] wrote:
 I really support your idea for a lod cloud that's actually useful to
 write queries, i promise we'll do the best from sindice to deliver one
 such a thing.

That sounds like a great idea for a service. Right now I just need a
little diagram that illustrates the links between resources in the LOD
cloud; and the diversity of descriptions from each provider. But, I
guess I should stop begging on here, and just create it eh?

Thanks for the feedback,
//Ed



Re: Drilling into the LOD Cloud

2008-09-22 Thread Ed Summers

Yep, that's the plan .. just wondering if someone had some slideware
content to that effect already :-)

//Ed

On Mon, Sep 22, 2008 at 4:42 PM, Kingsley Idehen [EMAIL PROTECTED] wrote:
 Ed Summers wrote:

 On Mon, Sep 22, 2008 at 1:28 PM, Giovanni Tummarello
 [EMAIL PROTECTED] wrote:


 I really support your idea for a lod cloud that's actually useful to
 write queries, i promise we'll do the best from sindice to deliver one
 such a thing.


 That sounds like a great idea for a service. Right now I just need a
 little diagram that illustrates the links between resources in the LOD
 cloud; and the diversity of descriptions from each provider. But, I
 guess I should stop begging on here, and just create it eh?

 Thanks for the feedback,
 //Ed




 Ed,

 Are you not able to add additional rdf:type links between DBpedia and the
 following:

 1. Yago
 2. OpenCyc
 3. UMBEL

 Then between UMBEL and OpenCyc:

 1. owl:sameAs
 2. owl:equivalentClass


 Then between OpenCyc and Wordnet:
 1. owl:sameAs


 --


 Regards,

 Kingsley Idehen   Weblog: http://www.openlinksw.com/blog/~kidehen
 President  CEO OpenLink Software Web: http://www.openlinksw.com








Library of Congress Subject Headings as SKOS Linked Data

2008-06-09 Thread Ed Summers

I'd like to announce an experimental linked-data, SKOS representation
of the Library of Congress Subject Headings (LCSH) [1] ... and also
ask for some help.

The Library of Congress has been participating in the W3C Semantic Web
Deployment Working Group, and has converted LCSH from the MARC21 data
format [2] to SKOS. LCSH is a controlled vocabulary used to index
materials that have been added to the collections at the Library of
Congress. It has been in active development since 1898, and was first
published in 1914 so that other libraries and bibliographic utilities
could use and adapt it. The lcsh.info service makes 266,857 subject
headings available as SKOS concepts, which amounts to 2,441,494
triples that are separately downloadable [3] (since there isn't a
SPARQL endpoint just yet).

At the last SWDWG telecon some questions came up about the way
concepts are identified, and made available via HTTP. Since we're
hoping lcsh.info can serve as an implementation of SKOS for the W3C
recommendation process we want to make sure we do this right. So I was
hoping interested members of the linked-data and SKOS communities
could take a look and make sure the implementation looks correct.

Each concept is identified with a URI like:

  http://lcsh.info/sh95000541#concept

When responding to requests for concept URIs, the server content
negotiates to determine which representation of the concept to return:

  - application/xhtml+xml
  - application/json
  - text/n3
  - application/rdf+xml

This is basically the pattern that Cool URIs for the Semantic Web
discusses as the Hash URI with Content Negotiation [4]. An additional
point that is worth mentioning is that the XHTML representation
includes RDFa, that also describes the concept.

At the moment the LCSH/SKOS data is only linked to itself, through
assertions that involve skos:broader, skos:narrower, and skos:related.
But the hope is that minting URIs for LCSH will allow it to be mapped
and/or linked to concepts in other vocabularies: dbpedia, geonames,
etc.

Any feedback, criticisms, ideas are welcome either on either the
public-lod [5] or public-swd-wg [6] discussion lists.

Thanks for reading this far!
//Ed

[1] http://lcsh.info
[2] http://www.loc.gov/marc/
[3] http://lcsh.info/static/lcsh.nt
[4] http://www.w3.org/TR/cooluris/#hashuri
[5] http://lists.w3.org/Archives/Public/public-lod/
[6] http://lists.w3.org/Archives/Public/public-swd-wg/