On 6/16/11 6:53 PM, Giovanni Tummarello wrote:
Hi Tim ,

"documents" per se (a la HTTP response 200 response) on the web are less and less relevant as opposed to the "conceptual entities" that are represented by this document and held e.g. as DB records inside CMS, social networks etc.

e.g. a social network is about "people" those are the important entities. Then there might be 1000 different HTTP documents that you can get e.g.i f you're logged if you're not logged, if you have a cookie if you have another cookie, if you add &format=print. Specific URLs are pretty irrelevant as they contain all sort of extra information.

Layouts of CMS or web apps change all the time (and so do the HTML docs) but not the entities.

that's why "http response 200" level annotations are of such little ambiguity really you say you have so many annotations about documents, i honestly dont understand what you're referring to, are these HTTP retrievable documents?

Tim is saying, and pretty clearly: there are a lot of resources in HTML format on the Web. You access these via URLs (Addresses). Basically, you GET data from Addresses.

where are the annotations? are we talking about the http headers? about the "meta" tags in the <head> these are about the subject of the page too most of the time, not the page itself.

In these resources (projected as HTML documents) there is a lot of metadata. Dig a little deeper, there are also varying degrees of metadata in the HTTP responses. What doesn't exist is use of an abstraction whereby the Subject Matter items (what the HTML docs are about) are Identified by Names that resolve to their Representations which are best served via graph pictorials.


and this is the idea behind schema.org <http://schema.org> (opengraph whatever) which sorry Tim you have to live with and we have to do the most with.
Tim is not saying he has a problem with schema.org. He might imply that schema.org is deviating from the aspect of AWWW that delivers the abstraction necessary for schema.org to refer to entities (data objects) by Names that are distinct from the Addresses of their Representation(s).


When someone refers to a URL which embeds a opengraph or schema.org <http://schema.org> annotation then it is 99.+ (with the number of 9 augmenting as the web evolves to a rich app platform) certain that they refer to the entity described in it and not to the web document itself (which can and does change all the time and is of overall no conceptual relevance).

We are already dealing with the schema.org issues [1][2] the best way it can be handled until opportunity costs veer them towards upping the semantic fidelity of their contribution.

We can live with schema.org, but lets not conflate that effort with some vital fundamentals re. Linked Data and best practices based on AWWW. In my eyes, schema.org is a massive vector re. structured data injected into the Web. Semantic fidelity was never their focus. Basically, as stated in an older post, Google, Microsoft, Yahoo!, Facebook and friends, all seek to contribute structured data from their respective data spaces, this already makes sense to them and is 100% compatible with their respective business models. Naturally, we would like them to do more, but you can't tell them to do more, all you can do is make opportunity costs palpable to them and eventually they will respond.


With respect to schema.org <http://schema.org>, we (as semantic web community) have not been ignored: our work and proposals have been very well considered and then diregarded alltogether - and for several reasons : 12 years of work, not an agreement on ontology, not an easy way for people to publish data ( the 303 thing is a complete total utter insanity (as i had said in vain so many times) ). etc.

303 isn't insanity! It is basic computing re. data access by reference. de-reference (indirection) and address-of operations are fundamental elements of any kind of environment that allows access, movement, and manipulation of data. You always have Names and Addresses. In fact, you have them in the real world, but I don't want veer down a discussion on semiotics and philosophy.

The Web of Documents works because Document Addresses (URLs) have become intuitive. Evolving the Web to a Linked Data Space is a little trickier with HTTP URIs because the Name operation unveils a powerful but unnatural abstraction due to the fact that an HTTP URI based Name looks and feels like an HTTP URI based Location Name (Address).

HTTP 303 is just doing what programming languages do behind the scenes whenever you access data objects by reference. If anything, its a tribute to the flexibility of the HTTP protocol. Basically, we have the Web now pulling of the same data access and manipulation capabilities that host operating systems have delivered to systems developers since forever.

[SNIP]

Kingsley

Gio








On Thu, Jun 16, 2011 at 7:04 PM, Tim Berners-Lee <[email protected] <mailto:[email protected]>> wrote:

    I disagree with this post very strongly, and it is hard to know
    where to start,
    and I am surprised to see it.

    On 2011-06 -13, at 07:41, Richard Cyganiak wrote:

    > On 13 Jun 2011, at 09:59, Christopher Gutteridge wrote:
    >> The real problem seems to me that making resolvable, HTTP URIs
    for real world things was a clever but dirty hack and does not
    make any semantic sense.
    >
    > Well, you worry about *real-world things*, but even people who
    just worry about *documents* have said for two decades that the
    web is broken because it conflates names and addresses.

    No, some people didn't get the architecture in that they had
    learned systems where there that
    there was a big distinction between names and address, and they
    had different properties,
    and then they came across URIs which had properties of both.


    > And they keep proposing things like URNs and info: URIs and tag:
    URIs and XRIs and DOIs to fix that and to separate the naming
    concern from the address concern. And invariably, these things
    fizzle around in their little niche for a while and then mostly
    die, because this aspect that you call a “clever but dirty hack”
    is just SO INCREDIBLY USEFUL. And being useful trumps making
    semantic sense.

    I agree ... except that ther URI architectre being like names and
    like addresses isn't a "clever but dirty hack".

    You then connect this with the idea of using HTTP URIs for
    real-world things, which is a separate queston.
    This again is a question of architecture. Of design of a system.
    We can make it work either way.
    We have to work out which is best.

    I don't think 303 is a quick and dirty hack.
    It does mean a large extension of HTTP to be uses with non-documents.
    It does have efficiency problems.
    It is an architectural extension to the web architecture.

    >
    > HTTP has been successfully conflating names and addresses since
    1989.

    That is COMPLETELY irrelevant.
    It is not a question of the web being fuzzy or ambiguous and
    getting away with it.
    It is a clean architecture where the concepts of "name" and
    "address" don't connect directly with those of people or files on
    a disk or IP hosts.


    >
    > There is a trillion web pages out there, all named with URIs.
    And even if just 0.1% of these pages are unambiguously about a
    single specific thing, that gives us a billion free identifiers
    for real-world entities, all already equipped with rich
    *human-readable* representations, and already linked and
    interconnected with *human-readable*, untyped, @href links.
    >
    > And these one billion URIs are plain old http:// URIs. They
    don't have a thing:// in the beginning, nor a tdb://, nor a #this
    or #that in the end, nor do they respond with 303 redirects or to
    MGET requests or whatever other nutty proposals we have come up
    with over the years to disambiguate between page and topic. They
    are plain old http:// URIs. A billion.
    >
    > Then add to that another huge number that already responds with
    JSON or XML descriptions of some interesting entity, like the one
    from Facebook that Kingsley mentioned today in a parallel thread.
    Again, no thing:// or tdb:// or #this or 303 or MGET on any of them.
    >
    > I want to use these URIs as identifiers in my data, and I have
    no intention of redirecting through an intermediate blank node
    just because the TAG fucked up some years ago.

    If you want to give yourself the luxury of being able to refer to
    the subject of a webpage, without having to add anthing to
    disambiguate it from the web page, then for the sake of your
    system, so you can use the billion web pages for your purposes,
    then you now stop other like me from using semantic web systems to
    refer to those web pages, or in fact to the other hundred million
    web pages either.

    Maybe you should an efficient way of doing what you want without
    destroying the system (which you as well have done so much to build)



    >
    > I want to tell the publishers of these web pages that they could
    join the web of data just by adding a few @rels to some <a>s, and
    a few @properties to some <span>s, and a few @typeofs to some
    <div>s (or @itemtypes and @itemprops). And I don't want to explain
    to them that they should also change http:// to thing:// or tdb://
    or add #this or #that or make their stuff respond with 303 or to
    MGET requests because you can't squeeze a dog through an HTTP
    connection.

    Well actually I really want them to put metadata about BOTH the
    document and its subject.

    There is masses of metadata already about documents.

    Now you want to make it ambiguous so I don't know whether it is
    about the document or its subject?

    I don't think something like about="#product" is rocket science or
    unnatural.

    I really want people to be able to use RDF or microdata to say
    things about more than one thing in the same page

    >
    > And here you and Pat and Alan (and TimBL, for that matter) are
    preaching that we can't use this one billion of fantastic free
    URIs to identify things because it wouldn't make semantic sense.

    We are saying that actually we already are using them to refer to
    the web pages and that that is very important and so is all the
    existing web.

    >
    > Being useful trumps making semantic sense.

    That is romantic nonsense.  To be useful you need clean extensible
    architecture,
    well defined concepts.

    > The web succeeded *because* it conflates name and address.

    That is completely irrelevant nonsense.


    It succeeded with a clean architecture using URIs for web pages,
    and the # as punctuation syntax between the identifier of the page
    and the local identifier within the page.


    > The web of data will succeed *because* it conflates a thing and
    a web page about the thing.
    >
    > <http://richard.cyganiak.de/>
    >    a foaf:Document;
    >    dc:title "Richard Cyganiak's homepage";
    >    a foaf:Person;
    >    foaf:name "Richard Cyganiak";
    >    owl:sameAs <http://twitter.com/cygri>;
    >    .
    >
    > There.
    >
    > If your knowledge representation formalism isn't smart enough to
    make sense of that, then it may just not be quite ready for the
    web, and you may have some work to do.

    Formalisms aren't smart.
    Sure, I can make a program to make sense of that.
    But I'm not going to just to save you the effort of getting it right.

    Disappointed by the intensity of your posting.
    Systems have managed for a long time to distinguish between
    library car and book,
    between message header and message,
    between a book and its subject.

    Now we have masses of information about many books
    and about many other things we have great value in it
    Let's not mess it up.

    If you want an ambiguous source of information, use natural language.
    The power of data is that is a whole lot less ambiguous.

    Tim

    >
    > Best,
    > Richard
    >





--

Regards,

Kingsley Idehen 
President&  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen





Reply via email to