Sebastian, all:
The community may not publicly admit it, but: SW and LOD have been BEGGING for 
adoption for almost a decade. Now, if someone outside of a University project 
publishes valuable RDF data in a well-above-the-standards way, you make him pay 
several hundred Euros for traffic just for your ISWC paper.

Quote from one e-mail I received: "I think we are among the Universities 
running such a crawler. Because we were in a rush for the ISWC deadline, nobody 
took the time to implement robots.txt and bandwidth throttling. Sorry."

Stop dreaming. A technical improvement for the WWW cannot be developed in 
isolation from the socio-economic environment. I.e., it will lead to nowhere to 
just work on technical solutions that don't fit the characteristics of the 
target eco-system, skill-wise, incentive-wise, or complexity-wise, and then 
wait for the world to pick it up. Unless you want your work to be listed here

    http://www.mclol.com/funny-articles/most-useless-inventions-ever/

WebID is a notable exception, because it takes into account exactly those 
dimensions.

> And what if in the future 100.000 software agents will access servers? We 
> will have the scalability issue eventually even without crawlers, so let's 
> try to solve it. In the eyeball web, there are also crawlers without too much 
> of a problem, and if Linked Data is to be successful we need to do the same.

How do you personally solve the scalability issue for small site-owners who are 
running a decent service from a basic understanding of HTML, PHP, and MySQL?

Best
Martin


On Jun 23, 2011, at 1:08 AM, Sebastian Schaffert wrote:

> 
> Am 22.06.2011 um 23:01 schrieb Lin Clark:
> 
>> On Wed, Jun 22, 2011 at 9:33 PM, Sebastian Schaffert 
>> <[email protected]> wrote:
>> 
>> Your complaint sounds to me a bit like "help, too many clients access my 
>> data".
>> 
>> I'm sure that Martin is really tired of saying this, so I will reiterate for 
>> him: It wasn't his data, they weren't his servers. He's speaking on behalf 
>> of people who aren't part of our insular community... people who don't have 
>> a compelling reason to subsidize a PhD student's Best Paper award with their 
>> own dollars and bandwidth.
> 
> And what about those companies subsidizing PhD students who write crawlers 
> for the normal Web? Like Larry Page in 1998?
> 
>> 
>> Agents can use Linked Data just fine without firing 150 requests per second 
>> at a server. There are TONs of use cases that do not require that kind of 
>> server load.
> 
> And what if in the future 100.000 software agents will access servers? We 
> will have the scalability issue eventually even without crawlers, so let's 
> try to solve it. In the eyeball web, there are also crawlers without too much 
> of a problem, and if Linked Data is to be successful we need to do the same.
> 
> Greetings,
> 
> Sebastian
> -- 
> | Dr. Sebastian Schaffert          [email protected]
> | Salzburg Research Forschungsgesellschaft  http://www.salzburgresearch.at
> | Head of Knowledge and Media Technologies Group          +43 662 2288 423
> | Jakob-Haringer Strasse 5/II
> | A-5020 Salzburg
> 


Reply via email to