Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Steve Harris
While I don't agree with Andreas exactly that it's the site owners fault, this 
is something that publishers of non-semantic data have to deal with.

If you publish a large collection of interlinked data which looks interesting 
to conventional crawlers and is expensive to generate, conventional web 
crawlers will be all over it. The main difference is that a greater percentage 
of those are written properly, to follow robots.txt and the guidelines about 
hit frequency (maximum 1 request per second per domain, no parallel crawling).

Has someone published similar guidelines for semantic web crawlers?

The ones that don't behave themselves get banned, either in robots.txt, or 
explicitly by the server. 

- Steve

On 2011-06-22, at 06:07, Martin Hepp wrote:

 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are to 
 blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit by 
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF datasets 
 to protect themselves against hungry crawlers and actual DoS attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting 
 Semantic Sitemap information for DoS attacks, focussing on the large dump 
 files. 
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your servers 
 and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from 
 site-owners that Semantic Web crawlers from Universities visited their 
 sites in a way close to a denial-of-service attack, i.e., crawling data 
 with maximum bandwidth in a parallelized approach.
 
 It's clear that a single, stupidly written crawler script, run from a 
 powerful University network, can quickly create terrible traffic load. 
 
 Many of the scripts we saw
 
 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or lacked 
 contact information therein, 
 - used no mechanisms at all for limiting the default crawling speed and 
 re-crawling delays.
 
 This irresponsible behavior can be the final reason for site-owners to say 
 farewell to academic/W3C-sponsored semantic technology.
 
 So please, please - advise all of your colleagues and students to NOT write 
 simple crawler scripts for the billion triples challenge or whatsoever 
 without familiarizing themselves with the state of the art in friendly 
 crawling.
 
 Best wishes
 
 Martin Hepp
 
 
 
 

-- 
Steve Harris, CTO, Garlik Limited
1-3 Halford Road, Richmond, TW10 6AW, UK
+44 20 8439 8203  http://www.garlik.com/
Registered in England and Wales 535 7233 VAT # 849 0517 11
Registered office: Thames House, Portsmouth Road, Esher, Surrey, KT10 9AD




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Yves Raimond
Hello!

 The difference between these two scenarios is that there's almost no CPU
 involvement in serving the PDF file, but naive RDF sites use lots of cycles
 to generate the response to a query for an RDF document.

 Right now queries to data.southampton.ac.uk (eg.
 http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made
 live, but this is not efficient. My colleague, Dave Challis, has prepared a
 SPARQL endpoint which caches results which we can turn on if the load gets
 too high, which should at least mitigate the problem. Very few datasets
 change in a 24 hours period.

Hmm, I would strongly argue it is not the case (and stale datasets are
a bit issue in LOD imho!). The data on the BBC website, for example,
changes approximately 10 times a second.

We've also been hit in the past (and still now, to a lesser extent) by
badly behaving crawlers. I agree that, as we don't provide dumps, it
is the only way to generate an aggregation of BBC data, but we've had
downtime in the past caused by crawlers. After that happened, it
caused lots of discussions on whether we should publish RDF data at
all (thankfully, we succeeded to argue that we should keep it - but
that's a lot of time spent arguing instead of publishing new juicy RDF
data!)

I also want to point out (in response to Andreas's email) that HTTP
caches are *completely* inefficient to protect a dataset against that,
as crawlers tend to be exhaustive. ETags and Expiry headers are
helpful, but chances are that 1) you don't know when the data will
change, you can just make a wild guess based on previous behavior 2)
the cache would have expired the time the crawler requests a document
a second time, as it has ~100M (in our case) documents to crawl
through.

Request throttling would work, but you would have to find a way to
identify crawlers, which is tricky: most of them use multiple IPs and
don't set appropriate user agents (the crawlers that currently hit us
the most are wget and Java 1.6 :/ ).

So overall, there is no excuse for badly behaving crawlers!

Cheers,
y


 Martin Hepp wrote:

 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.

 However, I still challenge Andreas' statement that the site-owners are to
 blame for publishing large amounts of data on small servers.

 One can publish 10,000 PDF documents on a tiny server without being hit by
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?

 But for sure, it is necessary to advise all publishers of large RDF datasets
 to protect themselves against hungry crawlers and actual DoS attacks.

 Imagine if a large site was brought down by a botnet that is exploiting
 Semantic Sitemap information for DoS attacks, focussing on the large dump
 files.
 This could end LOD experiments for that site.


 Best

 Martin


 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:



 Hi Martin,

 Have you tried to put a Squid [1]  as reverse proxy in front of your servers
 and use delay pools [2] to catch hungry crawlers?

 Cheers,
 Daniel

 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools

 On 21.06.2011, at 09:49, Martin Hepp wrote:



 Hi all:

 For the third time in a few weeks, we had massive complaints from
 site-owners that Semantic Web crawlers from Universities visited their sites
 in a way close to a denial-of-service attack, i.e., crawling data with
 maximum bandwidth in a parallelized approach.

 It's clear that a single, stupidly written crawler script, run from a
 powerful University network, can quickly create terrible traffic load.

 Many of the scripts we saw

 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or lacked
 contact information therein,
 - used no mechanisms at all for limiting the default crawling speed and
 re-crawling delays.

 This irresponsible behavior can be the final reason for site-owners to say
 farewell to academic/W3C-sponsored semantic technology.

 So please, please - advise all of your colleagues and students to NOT write
 simple crawler scripts for the billion triples challenge or whatsoever
 without familiarizing themselves with the state of the art in friendly
 crawling.

 Best wishes

 Martin Hepp





 --
 Christopher Gutteridge -- http://id.ecs.soton.ac.uk/person/1248

 You should read the ECS Web Team blog: http://blogs.ecs.soton.ac.uk/webteam/




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Martin Hepp
Just to inform the community that the BTC / research crawlers have been 
successful in killing a major RDF source for e-commerce:

OpenEAN - a transcript of 1 Mio product models and their EAN/UPC code at 
http://openean.kaufkauf.net/id/ has been permanently shut down by the site 
operator because fighting with bad semweb crawlers is taking too much of his 
time.

Thanks a lot for everybody who contributed to that. It trashes a month of work 
and many million useful triples.

Best

Martin Hepp



On Jun 22, 2011, at 11:37 AM, Yves Raimond wrote:

 Hello!
 
 The difference between these two scenarios is that there's almost no CPU
 involvement in serving the PDF file, but naive RDF sites use lots of cycles
 to generate the response to a query for an RDF document.
 
 Right now queries to data.southampton.ac.uk (eg.
 http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made
 live, but this is not efficient. My colleague, Dave Challis, has prepared a
 SPARQL endpoint which caches results which we can turn on if the load gets
 too high, which should at least mitigate the problem. Very few datasets
 change in a 24 hours period.
 
 Hmm, I would strongly argue it is not the case (and stale datasets are
 a bit issue in LOD imho!). The data on the BBC website, for example,
 changes approximately 10 times a second.
 
 We've also been hit in the past (and still now, to a lesser extent) by
 badly behaving crawlers. I agree that, as we don't provide dumps, it
 is the only way to generate an aggregation of BBC data, but we've had
 downtime in the past caused by crawlers. After that happened, it
 caused lots of discussions on whether we should publish RDF data at
 all (thankfully, we succeeded to argue that we should keep it - but
 that's a lot of time spent arguing instead of publishing new juicy RDF
 data!)
 
 I also want to point out (in response to Andreas's email) that HTTP
 caches are *completely* inefficient to protect a dataset against that,
 as crawlers tend to be exhaustive. ETags and Expiry headers are
 helpful, but chances are that 1) you don't know when the data will
 change, you can just make a wild guess based on previous behavior 2)
 the cache would have expired the time the crawler requests a document
 a second time, as it has ~100M (in our case) documents to crawl
 through.
 
 Request throttling would work, but you would have to find a way to
 identify crawlers, which is tricky: most of them use multiple IPs and
 don't set appropriate user agents (the crawlers that currently hit us
 the most are wget and Java 1.6 :/ ).
 
 So overall, there is no excuse for badly behaving crawlers!
 
 Cheers,
 y
 
 
 Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are to
 blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit by
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF datasets
 to protect themselves against hungry crawlers and actual DoS attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting
 Semantic Sitemap information for DoS attacks, focussing on the large dump
 files.
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your servers
 and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from
 site-owners that Semantic Web crawlers from Universities visited their sites
 in a way close to a denial-of-service attack, i.e., crawling data with
 maximum bandwidth in a parallelized approach.
 
 It's clear that a single, stupidly written crawler script, run from a
 powerful University network, can quickly create terrible traffic load.
 
 Many of the scripts we saw
 
 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or lacked
 contact information therein,
 - used no mechanisms at all for limiting the default crawling speed and
 re-crawling delays.
 
 This irresponsible behavior can be the final reason for site-owners to say
 farewell to academic/W3C-sponsored semantic technology.
 
 So please, please - advise all of your colleagues and students to NOT write
 simple crawler scripts for the billion triples challenge or whatsoever
 without familiarizing themselves with the state of the art in friendly
 crawling.
 
 Best wishes
 
 Martin Hepp
 
 
 
 
 
 --
 Christopher Gutteridge -- 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Andreas Harth

Hi Christopher,

On 06/22/2011 10:14 AM, Christopher Gutteridge wrote:

Right now queries to data.southampton.ac.uk (eg.
http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made live,
but this is not efficient. My colleague, Dave Challis, has prepared a SPARQL
endpoint which caches results which we can turn on if the load gets too high,
which should at least mitigate the problem. Very few datasets change in a 24
hours period.


setting the Expires header and enabling mod_cache in Apache httpd (or adding
a Squid proxy in front of the HTTP server) works quite well in these cases.

Best regards,
Andreas.



Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 10:37 AM, Yves Raimond wrote:

Request throttling would work, but you would have to find a way to
identify crawlers, which is tricky: most of them use multiple IPs and
don't set appropriate user agents (the crawlers that currently hit us
the most are wget and Java 1.6 :/ ).
Hence the requirement for incorporation of WebID as basis for QoS for 
identifier agents. Everyone else gets to be constrained with rate limits 
etc..


Anyway, Identification is the key, the InterWeb jungle needs WebID to 
help reduce costs of serving up Linked Data etc..


Amazing its taken us until 2011 to revisit this critical matter.

--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Jiří Procházka
I wonder, are ways to link RDF data so that convential crawlers do not
crawl it, but only the semantic web aware ones do?
I am not sure how the current practice of linking by link tag in the
html headers could cause this, but it may be case that those heavy loads
come from a crawlers having nothing to do with semantic web...
Maybe we should start linking to our rdf/xml, turtle, ntriples files and
publishing sitemap info in RDFa...

Best,
Jiri

On 06/22/2011 09:00 AM, Steve Harris wrote:
 While I don't agree with Andreas exactly that it's the site owners fault, 
 this is something that publishers of non-semantic data have to deal with.
 
 If you publish a large collection of interlinked data which looks interesting 
 to conventional crawlers and is expensive to generate, conventional web 
 crawlers will be all over it. The main difference is that a greater 
 percentage of those are written properly, to follow robots.txt and the 
 guidelines about hit frequency (maximum 1 request per second per domain, no 
 parallel crawling).
 
 Has someone published similar guidelines for semantic web crawlers?
 
 The ones that don't behave themselves get banned, either in robots.txt, or 
 explicitly by the server. 
 
 - Steve
 
 On 2011-06-22, at 06:07, Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.

 However, I still challenge Andreas' statement that the site-owners are to 
 blame for publishing large amounts of data on small servers.

 One can publish 10,000 PDF documents on a tiny server without being hit by 
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?

 But for sure, it is necessary to advise all publishers of large RDF datasets 
 to protect themselves against hungry crawlers and actual DoS attacks.

 Imagine if a large site was brought down by a botnet that is exploiting 
 Semantic Sitemap information for DoS attacks, focussing on the large dump 
 files. 
 This could end LOD experiments for that site.


 Best

 Martin


 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:


 Hi Martin,

 Have you tried to put a Squid [1]  as reverse proxy in front of your 
 servers and use delay pools [2] to catch hungry crawlers?

 Cheers,
 Daniel

 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools

 On 21.06.2011, at 09:49, Martin Hepp wrote:

 Hi all:

 For the third time in a few weeks, we had massive complaints from 
 site-owners that Semantic Web crawlers from Universities visited their 
 sites in a way close to a denial-of-service attack, i.e., crawling data 
 with maximum bandwidth in a parallelized approach.

 It's clear that a single, stupidly written crawler script, run from a 
 powerful University network, can quickly create terrible traffic load. 

 Many of the scripts we saw

 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or 
 lacked contact information therein, 
 - used no mechanisms at all for limiting the default crawling speed and 
 re-crawling delays.

 This irresponsible behavior can be the final reason for site-owners to say 
 farewell to academic/W3C-sponsored semantic technology.

 So please, please - advise all of your colleagues and students to NOT 
 write simple crawler scripts for the billion triples challenge or 
 whatsoever without familiarizing themselves with the state of the art in 
 friendly crawling.

 Best wishes

 Martin Hepp




 



signature.asc
Description: OpenPGP digital signature


Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 10:42 AM, Martin Hepp wrote:

Just to inform the community that the BTC / research crawlers have been 
successful in killing a major RDF source for e-commerce:

OpenEAN - a transcript of1 Mio product models and their EAN/UPC code at 
http://openean.kaufkauf.net/id/ has been permanently shut down by the site 
operator because fighting with bad semweb crawlers is taking too much of his time.

Thanks a lot for everybody who contributed to that. It trashes a month of work 
and many million useful triples.


Martin,

Is there a dump anywhere? Can they at least continue to produce RDF dumps?

We have some of their data (from prior dump loads) in our lod cloud 
cache [1].


Links:

1. 
http://lod.openlinksw.com/describe/?url=http%3A%2F%2Fopenean.kaufkauf.net%2Fid%2Furilookup=1



Kingsley

Best

Martin Hepp



On Jun 22, 2011, at 11:37 AM, Yves Raimond wrote:


Hello!


The difference between these two scenarios is that there's almost no CPU
involvement in serving the PDF file, but naive RDF sites use lots of cycles
to generate the response to a query for an RDF document.

Right now queries to data.southampton.ac.uk (eg.
http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made
live, but this is not efficient. My colleague, Dave Challis, has prepared a
SPARQL endpoint which caches results which we can turn on if the load gets
too high, which should at least mitigate the problem. Very few datasets
change in a 24 hours period.

Hmm, I would strongly argue it is not the case (and stale datasets are
a bit issue in LOD imho!). The data on the BBC website, for example,
changes approximately 10 times a second.

We've also been hit in the past (and still now, to a lesser extent) by
badly behaving crawlers. I agree that, as we don't provide dumps, it
is the only way to generate an aggregation of BBC data, but we've had
downtime in the past caused by crawlers. After that happened, it
caused lots of discussions on whether we should publish RDF data at
all (thankfully, we succeeded to argue that we should keep it - but
that's a lot of time spent arguing instead of publishing new juicy RDF
data!)

I also want to point out (in response to Andreas's email) that HTTP
caches are *completely* inefficient to protect a dataset against that,
as crawlers tend to be exhaustive. ETags and Expiry headers are
helpful, but chances are that 1) you don't know when the data will
change, you can just make a wild guess based on previous behavior 2)
the cache would have expired the time the crawler requests a document
a second time, as it has ~100M (in our case) documents to crawl
through.

Request throttling would work, but you would have to find a way to
identify crawlers, which is tricky: most of them use multiple IPs and
don't set appropriate user agents (the crawlers that currently hit us
the most are wget and Java 1.6 :/ ).

So overall, there is no excuse for badly behaving crawlers!

Cheers,
y


Martin Hepp wrote:

Hi Daniel,
Thanks for the link! I will relay this to relevant site-owners.

However, I still challenge Andreas' statement that the site-owners are to
blame for publishing large amounts of data on small servers.

One can publish 10,000 PDF documents on a tiny server without being hit by
DoS-style crazy crawlers. Why should the same not hold if I publish RDF?

But for sure, it is necessary to advise all publishers of large RDF datasets
to protect themselves against hungry crawlers and actual DoS attacks.

Imagine if a large site was brought down by a botnet that is exploiting
Semantic Sitemap information for DoS attacks, focussing on the large dump
files.
This could end LOD experiments for that site.


Best

Martin


On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:



Hi Martin,

Have you tried to put a Squid [1]  as reverse proxy in front of your servers
and use delay pools [2] to catch hungry crawlers?

Cheers,
Daniel

[1] http://www.squid-cache.org/
[2] http://wiki.squid-cache.org/Features/DelayPools

On 21.06.2011, at 09:49, Martin Hepp wrote:



Hi all:

For the third time in a few weeks, we had massive complaints from
site-owners that Semantic Web crawlers from Universities visited their sites
in a way close to a denial-of-service attack, i.e., crawling data with
maximum bandwidth in a parallelized approach.

It's clear that a single, stupidly written crawler script, run from a
powerful University network, can quickly create terrible traffic load.

Many of the scripts we saw

- ignored robots.txt,
- ignored clear crawling speed limitations in robots.txt,
- did not identify themselves properly in the HTTP request header or lacked
contact information therein,
- used no mechanisms at all for limiting the default crawling speed and
re-crawling delays.

This irresponsible behavior can be the final reason for site-owners to say
farewell to academic/W3C-sponsored semantic technology.

So please, please - advise all of your colleagues and students to NOT write
simple crawler scripts for the billion 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread adam . saltiel
Yes. But are there things such as Squid and WebId that can be instituted the 
provider side? This is an interesting moment. Is it the academic SemWeb running 
out of public facing steam. A retreat. Or is it a moment of transition from 
naivety to responsibility. When we think about the Cathedral and the Bazaar. 
There is a reason why people want Cathedrals. I suggest SemWeb is about 
Cathedrals. Responsibility for some order and structure. 

Adam 
Sent using BlackBerry® from Orange

-Original Message-
From: Martin Hepp martin.h...@ebusiness-unibw.org
Sender: semantic-web-requ...@w3.org
Date: Wed, 22 Jun 2011 11:42:58 
To: Yves Raimondyves.raim...@gmail.com
Cc: Christopher Gutteridgec...@ecs.soton.ac.uk; Daniel 
Herzigher...@kit.edu; semantic-...@w3.org; public-lod@w3.org
Subject: Re: Think before you write Semantic Web crawlers

Just to inform the community that the BTC / research crawlers have been 
successful in killing a major RDF source for e-commerce:

OpenEAN - a transcript of 1 Mio product models and their EAN/UPC code at 
http://openean.kaufkauf.net/id/ has been permanently shut down by the site 
operator because fighting with bad semweb crawlers is taking too much of his 
time.

Thanks a lot for everybody who contributed to that. It trashes a month of work 
and many million useful triples.

Best

Martin Hepp



On Jun 22, 2011, at 11:37 AM, Yves Raimond wrote:

 Hello!
 
 The difference between these two scenarios is that there's almost no CPU
 involvement in serving the PDF file, but naive RDF sites use lots of cycles
 to generate the response to a query for an RDF document.
 
 Right now queries to data.southampton.ac.uk (eg.
 http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made
 live, but this is not efficient. My colleague, Dave Challis, has prepared a
 SPARQL endpoint which caches results which we can turn on if the load gets
 too high, which should at least mitigate the problem. Very few datasets
 change in a 24 hours period.
 
 Hmm, I would strongly argue it is not the case (and stale datasets are
 a bit issue in LOD imho!). The data on the BBC website, for example,
 changes approximately 10 times a second.
 
 We've also been hit in the past (and still now, to a lesser extent) by
 badly behaving crawlers. I agree that, as we don't provide dumps, it
 is the only way to generate an aggregation of BBC data, but we've had
 downtime in the past caused by crawlers. After that happened, it
 caused lots of discussions on whether we should publish RDF data at
 all (thankfully, we succeeded to argue that we should keep it - but
 that's a lot of time spent arguing instead of publishing new juicy RDF
 data!)
 
 I also want to point out (in response to Andreas's email) that HTTP
 caches are *completely* inefficient to protect a dataset against that,
 as crawlers tend to be exhaustive. ETags and Expiry headers are
 helpful, but chances are that 1) you don't know when the data will
 change, you can just make a wild guess based on previous behavior 2)
 the cache would have expired the time the crawler requests a document
 a second time, as it has ~100M (in our case) documents to crawl
 through.
 
 Request throttling would work, but you would have to find a way to
 identify crawlers, which is tricky: most of them use multiple IPs and
 don't set appropriate user agents (the crawlers that currently hit us
 the most are wget and Java 1.6 :/ ).
 
 So overall, there is no excuse for badly behaving crawlers!
 
 Cheers,
 y
 
 
 Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are to
 blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit by
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF datasets
 to protect themselves against hungry crawlers and actual DoS attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting
 Semantic Sitemap information for DoS attacks, focussing on the large dump
 files.
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your servers
 and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from
 site-owners that Semantic Web crawlers from Universities visited their sites
 in a way close to a denial-of-service attack, i.e., crawling data with
 maximum bandwidth in a parallelized approach.
 
 It's clear that a single, stupidly written crawler script, run from a
 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Lin Clark
I was with you on this until the cathedral and the bazaar thing... I think
it is a serious misreading of cathedral and bazaar to think that if
something is naive and irresponsible, it is by definition bazaar style
development. Bazaar style is about how code is developed (in the open by a
loosely organized and fluctuating group of developers). Cathedral means that
it is a smaller, generally hierarchically organized group which doesn't work
in a public, open way between releases. There is a good summary on
wikipedia, http://en.wikipedia.org/wiki/The_Cathedral_and_the_Bazaar

The bazaar style of development can lead to things that are more responsible
than their cathedral counterparts. Bazaar means continuously documenting
your decisions in the public, posting patches for testing and review by
everyone (not just your extremely busy team mates), and opening your dev
process to co-developers who you don't already know. These organizational
strategies have lead to some REALLY BIG engineering wins... and these
engineering wins have resulted in more responsible products than their
cathedral-built counterparts.

I also would question the assertion that people want cathedrals... the
general direction on the Web seems to be away from cathedrals like Microsoft
and Flash and towards bazaar developed solutions.

However, the call to responsibility is still a very valid one. I'm quite
sorry to hear that a large data publisher has been pushed out of the
community effort by people who should be working on the same team.

-Lin


On Wed, Jun 22, 2011 at 11:59 AM, adam.salt...@gmail.com wrote:

 Yes. But are there things such as Squid and WebId that can be instituted
 the provider side? This is an interesting moment. Is it the academic SemWeb
 running out of public facing steam. A retreat. Or is it a moment of
 transition from naivety to responsibility. When we think about the Cathedral
 and the Bazaar. There is a reason why people want Cathedrals. I suggest
 SemWeb is about Cathedrals. Responsibility for some order and structure.

 Adam
 Sent using BlackBerry® from Orange

 -Original Message-
 From: Martin Hepp martin.h...@ebusiness-unibw.org
 Sender: semantic-web-requ...@w3.org
 Date: Wed, 22 Jun 2011 11:42:58
 To: Yves Raimondyves.raim...@gmail.com
 Cc: Christopher Gutteridgec...@ecs.soton.ac.uk; Daniel Herzig
 her...@kit.edu; semantic-...@w3.org; public-lod@w3.org
 Subject: Re: Think before you write Semantic Web crawlers

 Just to inform the community that the BTC / research crawlers have been
 successful in killing a major RDF source for e-commerce:

 OpenEAN - a transcript of 1 Mio product models and their EAN/UPC code at
 http://openean.kaufkauf.net/id/ has been permanently shut down by the site
 operator because fighting with bad semweb crawlers is taking too much of his
 time.

 Thanks a lot for everybody who contributed to that. It trashes a month of
 work and many million useful triples.

 Best

 Martin Hepp



 On Jun 22, 2011, at 11:37 AM, Yves Raimond wrote:

  Hello!
 
  The difference between these two scenarios is that there's almost no CPU
  involvement in serving the PDF file, but naive RDF sites use lots of
 cycles
  to generate the response to a query for an RDF document.
 
  Right now queries to data.southampton.ac.uk (eg.
  http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are
 made
  live, but this is not efficient. My colleague, Dave Challis, has
 prepared a
  SPARQL endpoint which caches results which we can turn on if the load
 gets
  too high, which should at least mitigate the problem. Very few datasets
  change in a 24 hours period.
 
  Hmm, I would strongly argue it is not the case (and stale datasets are
  a bit issue in LOD imho!). The data on the BBC website, for example,
  changes approximately 10 times a second.
 
  We've also been hit in the past (and still now, to a lesser extent) by
  badly behaving crawlers. I agree that, as we don't provide dumps, it
  is the only way to generate an aggregation of BBC data, but we've had
  downtime in the past caused by crawlers. After that happened, it
  caused lots of discussions on whether we should publish RDF data at
  all (thankfully, we succeeded to argue that we should keep it - but
  that's a lot of time spent arguing instead of publishing new juicy RDF
  data!)
 
  I also want to point out (in response to Andreas's email) that HTTP
  caches are *completely* inefficient to protect a dataset against that,
  as crawlers tend to be exhaustive. ETags and Expiry headers are
  helpful, but chances are that 1) you don't know when the data will
  change, you can just make a wild guess based on previous behavior 2)
  the cache would have expired the time the crawler requests a document
  a second time, as it has ~100M (in our case) documents to crawl
  through.
 
  Request throttling would work, but you would have to find a way to
  identify crawlers, which is tricky: most of them use multiple IPs and
  don't set appropriate 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Andreas Harth

Hi Martin,

first let me say that I do think crawlers should follow basic politeness
rules (contact info in User-Agent, adhere to the Robot Exclusion Protocol).

However, I am delighted that people actually start consuming Linked Data,
and we should encourage that.

On 06/22/2011 11:42 AM, Martin Hepp wrote:

OpenEAN - a transcript of1 Mio product models and their EAN/UPC code at
http://openean.kaufkauf.net/id/ has been permanently shut down by the site
operator because fighting with bad semweb crawlers is taking too much of his
time.


I've put a wrapper online [1] that provides RDF based on their API (which,
incidentally, currently does not seem to work either).

The wrapper does some caching and has a limit of one lookup every 8 seconds,
which means (24*60*60)/8 = 10800 lookups per day.  Data transfer is capped
to 1 GB/day, which means a maximum cost of 0.15 Euro/day at Amazon AWS pricing.

At that rate, it would take 925 days to collect descriptions of just one
million products.  Whether the ratio of data size and lookup limit is sensible
in that case is open to debate.

If the OpenEAN guys can redirect requests to [1] there would even be some
continuity for data consumers.

Best regards,
Andreas.

[1] http://openeanwrap.appspot.com/



Re: Hackers - Re: Schema.org considered helpful

2011-06-22 Thread adasal
Hi
I haven't had time to follow link
I expect there is an issue of how to think about a semantic web.
I can see Google is about ruthlessly exploiting the atomisation of the
Bazaar. Of course from within the walls of their own Cathedral.
Recall is in inverse proportion to accuracy.
I think web behaviours influence our own (mind) behaviours. We respond
to environment. Hints from that environment are assimilated very
quickly.
The web is an (absorbing for important reasons undiscussed here) environment.
I rely on Google very happily. It brings fragments some times random
often according to rules I half guess at. This is how it deals with
recall/accuracy.
SemWeb should be different. It is machine/machine. But there is an
ultimate human arbiter of relevance and quality of data for human
consumption. SemWeb needs a series of a priories - the ontologies.
It seems there are two human arbiter questions.
1. What data would I like to see - describe a coherent package of concepts.
2. Describe an ontology as a package of concepts.
In other words concept packages should be able to function independent
of attachment to ontology. And there needs a function to translate
between them. Ontology is already too low level.
It is impossible to characterise what people may be able to agree upon
as concept packages - data aims.
What people agree on depends on all the mixes of any human situation.
Is there a base strata of factors, a common field. I don't know but
I'm sure work has been done in the area. At simplest this is relation
between beliefs, hopes and desires which can never fully be known and
intersect in some group such that an agreed model can be made.
Models aspire to this. Groups create rules to facilitate this.
This is the responsibility the semweb has.
1. To identify such means of modelling and
2. mediate (show what it takes; what it is like to mediate) the
movement between model and some norms.
Here I mean behavioural norms. (So they need to be established case by
case. WebId to prevent unfriendly crawlers is a good simple example)
Not logical rules.
It is only with this in mind that anything of interest can be created.
Note: this is not creating something in the Bazaar of random market
forces. And, as with all heavily patterned behaviour, this is very
expensive in effort. It is also without the background data generation
of google as we traverse their graph. No gleaning off users. Radically
different.

Best

Adam

On 17/06/2011, Henry Story henry.st...@bblfish.net wrote:

 On 17 Jun 2011, at 19:27, adasal wrote:

 That said the hacker is a various beast,

 Indeed, hackers are not angels. But the people on this list should get back
 to hacking or work together with open source projects to get initial minimal
 working pieces embedded there. WebID is one; foaf is another, pingback,
 access control, ...
 Get the really simple pieces working.

 and I wonder if this sort of thing can really be addressed without
 overarching political/ethical/idealogical concerns. It's tough.

 It all fits together really nicely. I gave a talk on the philosophy of the
 Social Web if you are interested.
  http://www.slideshare.net/bblfish/philosophy-and-the-social-web-5583083

 Hackers tend to be engineers with a political attitude, so they are more
 receptive to the bigger picture. But solving the big picture problem should
 have an easy entry cost if we want to get it going.

 I talked to the BBC but they have limited themselves to what they will do in
 the Social Web space as far as profile hosting goes. Again, I'd start small.
 Facebook started in universities not that long ago.

 Henry


 Social Web Architect
 http://bblfish.net/



-- 
Sent from my mobile device



Re: Think before you write Semantic Web crawlers

2011-06-22 Thread glenn mcdonald
From my perspective as the designer of a system that both consumes and
publishes data, the load/burden issue here is not at all particular to the
semantic web. Needle obeys robots.txt rules, but that's a small deal
compared to the difficulty of extracting whole data from sites set up to
deliver it only in tiny pieces. I'd say about 98% of the time I can describe
the data I want from a site with a single conceptual query. Indeed, once
I've got the data into Needle I can almost always actually produce that
query. But on the source site, I usually can't, and thus we are forced to
waste everybody's time navigating the machines through superfluous
presentation rendering designed for people. 10-at-a-time results lists,
interminable AJAX refreshes, animated DIV reveals, grafting back together
the splintered bits of tree-traversals, etc. This is all absurdly
unnecessary. Why is anybody having to crawl an open semantic-web dataset?
Isn't there a download link, and/or a SPARQL endpoint? If there isn't, why
not? We're the Semantic Web, dammit. If we aren't the masters of data
interoperability, what are we?

glenn
(www.needlebase.com)


Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Martin Hepp
Thanks, Jiri, but the load comes from academic crawler prototypes firing from 
broad University infrastructures.
Best
Martin


On Jun 22, 2011, at 12:40 PM, Jiří Procházka wrote:

 I wonder, are ways to link RDF data so that convential crawlers do not
 crawl it, but only the semantic web aware ones do?
 I am not sure how the current practice of linking by link tag in the
 html headers could cause this, but it may be case that those heavy loads
 come from a crawlers having nothing to do with semantic web...
 Maybe we should start linking to our rdf/xml, turtle, ntriples files and
 publishing sitemap info in RDFa...
 
 Best,
 Jiri
 
 On 06/22/2011 09:00 AM, Steve Harris wrote:
 While I don't agree with Andreas exactly that it's the site owners fault, 
 this is something that publishers of non-semantic data have to deal with.
 
 If you publish a large collection of interlinked data which looks 
 interesting to conventional crawlers and is expensive to generate, 
 conventional web crawlers will be all over it. The main difference is that a 
 greater percentage of those are written properly, to follow robots.txt and 
 the guidelines about hit frequency (maximum 1 request per second per domain, 
 no parallel crawling).
 
 Has someone published similar guidelines for semantic web crawlers?
 
 The ones that don't behave themselves get banned, either in robots.txt, or 
 explicitly by the server. 
 
 - Steve
 
 On 2011-06-22, at 06:07, Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are to 
 blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit by 
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF 
 datasets to protect themselves against hungry crawlers and actual DoS 
 attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting 
 Semantic Sitemap information for DoS attacks, focussing on the large dump 
 files. 
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your 
 servers and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from 
 site-owners that Semantic Web crawlers from Universities visited their 
 sites in a way close to a denial-of-service attack, i.e., crawling data 
 with maximum bandwidth in a parallelized approach.
 
 It's clear that a single, stupidly written crawler script, run from a 
 powerful University network, can quickly create terrible traffic load. 
 
 Many of the scripts we saw
 
 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or 
 lacked contact information therein, 
 - used no mechanisms at all for limiting the default crawling speed and 
 re-crawling delays.
 
 This irresponsible behavior can be the final reason for site-owners to 
 say farewell to academic/W3C-sponsored semantic technology.
 
 So please, please - advise all of your colleagues and students to NOT 
 write simple crawler scripts for the billion triples challenge or 
 whatsoever without familiarizing themselves with the state of the art in 
 friendly crawling.
 
 Best wishes
 
 Martin Hepp
 
 
 
 
 
 




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 12:54 PM, Hugh Glaser wrote:

Hi Chris.
One way to do the caching really efficiently:
http://lists.w3.org/Archives/Public/semantic-web/2007Jun/0012.html
Which is what rkb has always done.
But of course caching does not solve the problem of one bad crawler.


Or a SPARQL query gone horribly wrong albeit inadvertently. In short, 
this is the most challenging case of all. We even had to protect against 
the same thing re. SQL access via ODBC, JDBC, ADO.NET, and OLE-DB. 
Basically, why we still have a business selling drivers when DBMS 
vendors offer free variants.




It actually makes it worse.
You add a cache write cost to the query, without a significant probability of a 
future cache hit. And increase disk usage.


Yes, and WebID adds fidelity to such inevitable challenges.

This is why (IMHO) WebID is the second most important innovation 
following the URI re., Linked Data.



Kingsley

Hugh

- Reply message -
From: Christopher Gutteridgec...@ecs.soton.ac.uk
To: Martin Heppmartin.h...@ebusiness-unibw.org
Cc: Daniel Herzigher...@kit.edu, semantic-...@w3.orgsemantic-...@w3.org, 
public-lod@w3.orgpublic-lod@w3.org
Subject: Think before you write Semantic Web crawlers
Date: Wed, Jun 22, 2011 9:18 am



The difference between these two scenarios is that there's almost no CPU 
involvement in serving the PDF file, but naive RDF sites use lots of cycles to 
generate the response to a query for an RDF document.

Right now queries to data.southampton.ac.uk (eg. 
http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made 
live, but this is not efficient. My colleague, Dave Challis, has prepared a 
SPARQL endpoint which caches results which we can turn on if the load gets too 
high, which should at least mitigate the problem. Very few datasets change in a 
24 hours period.

Martin Hepp wrote:

Hi Daniel,
Thanks for the link! I will relay this to relevant site-owners.

However, I still challenge Andreas' statement that the site-owners are to blame 
for publishing large amounts of data on small servers.

One can publish 10,000 PDF documents on a tiny server without being hit by 
DoS-style crazy crawlers. Why should the same not hold if I publish RDF?

But for sure, it is necessary to advise all publishers of large RDF datasets to 
protect themselves against hungry crawlers and actual DoS attacks.

Imagine if a large site was brought down by a botnet that is exploiting 
Semantic Sitemap information for DoS attacks, focussing on the large dump files.
This could end LOD experiments for that site.


Best

Martin


On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:



Hi Martin,

Have you tried to put a Squid [1]  as reverse proxy in front of your servers 
and use delay pools [2] to catch hungry crawlers?

Cheers,
Daniel

[1] http://www.squid-cache.org/
[2] http://wiki.squid-cache.org/Features/DelayPools

On 21.06.2011, at 09:49, Martin Hepp wrote:



Hi all:

For the third time in a few weeks, we had massive complaints from site-owners 
that Semantic Web crawlers from Universities visited their sites in a way close 
to a denial-of-service attack, i.e., crawling data with maximum bandwidth in a 
parallelized approach.

It's clear that a single, stupidly written crawler script, run from a powerful 
University network, can quickly create terrible traffic load.

Many of the scripts we saw

- ignored robots.txt,
- ignored clear crawling speed limitations in robots.txt,
- did not identify themselves properly in the HTTP request header or lacked 
contact information therein,
- used no mechanisms at all for limiting the default crawling speed and 
re-crawling delays.

This irresponsible behavior can be the final reason for site-owners to say 
farewell to academic/W3C-sponsored semantic technology.

So please, please - advise all of your colleagues and students to NOT write simple 
crawler scripts for the billion triples challenge or whatsoever without familiarizing 
themselves with the state of the art in friendly crawling.

Best wishes

Martin Hepp








--
Christopher Gutteridge -- http://id.ecs.soton.ac.uk/person/1248

You should read the ECS Web Team blog: http://blogs.ecs.soton.ac.uk/webteam/






--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Dieter Fensel

At 14:37 22.06.2011, Andreas Harth wrote:

Hi Martin,

first let me say that I do think crawlers should follow basic politeness
rules (contact info in User-Agent, adhere to the Robot Exclusion Protocol).

However, I am delighted that people actually start consuming Linked Data,
and we should encourage that.


The real challenge may be to achieve usage of data in a way that 
provide benefits to its provider.



--
Dieter Fensel
Director STI Innsbruck, University of Innsbruck, Austria
http://www.sti-innsbruck.at/
phone: +43-512-507-6488/5, fax: +43-512-507-9872




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 1:18 PM, Henry Story wrote:
You need to move to strong defences. This is what WebID provides very 
efficiently. Each resource can ask the requestor for their identity 
before giving access to a resource. It is completely decentralised and 
about as efficient as one can get.
So just as the power of computing has grown for everyone to write 
silly software so TLS and https has become cheaper and cheaper. Google 
is now moving to put all its servers behind https and so is Facebook. 
Soon all the web will be behind https - and that will massively 
increase the security on the whole web.


Increase security from snooping, yes-ish.

Only when you add WebID to the equation does TLS truly have an 
opportunity to be very smart.


The physics of computing has changed, genetic algorithms are going to 
become the norm rather than the exception. Trust Logics will separate 
the winners from the losers. This is an inevitability.


--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 3:00 PM, Dieter Fensel wrote:

At 14:37 22.06.2011, Andreas Harth wrote:

Hi Martin,

first let me say that I do think crawlers should follow basic politeness
rules (contact info in User-Agent, adhere to the Robot Exclusion 
Protocol).


However, I am delighted that people actually start consuming Linked 
Data,

and we should encourage that.


The real challenge may be to achieve usage of data in a way that 
provide benefits to its provider.





Closer, but the core algorithm or heuristic will always require 
something like WebID plus some serious Linked Data dog-fooding re. trust 
logics. Owl ! :-)



--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Karl Dubost

Le 21 juin 2011 à 03:49, Martin Hepp a écrit :
 Many of the scripts we saw
 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or lacked 
 contact information therein, 
 - used no mechanisms at all for limiting the default crawling speed and 
 re-crawling delays.


Do you have a list of those and how to identify them?
So we can put them in our blocking lists?

.htaccess or Apache config with rules such as:

# added for abusive downloads or not respecting robots.txt
SetEnvIfNoCase User-Agent .*Technorati*. bad_bot
SetEnvIfNoCase User-Agent .*WikioFeedBot*. bad_bot
# [… cut part of my list …]
Order Allow,Deny
Deny from 85.88.12.104
Deny from env=bad_bot
Allow from all



-- 
Karl Dubost - http://dev.opera.com/
Developer Relations  Tools, Opera Software




WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread William Waites
What does WebID have to do with JSON? They're somehow representative
of two competing trends.

The RDF/JSON, JSON-LD, etc. work is supposed to be about making it
easier to work with RDF for your average programmer, to remove the
need for complex parsers, etc. and generally to lower the barriers.

The WebID arrangement is about raising barriers. Not intended to be
the same kind of barriers, certainly the intent isn't to make
programmer's lives more difficult, rather to provide a good way to do
distributed authentication without falling into the traps of PKI and
such.

While I like WebID, and I think it is very elegant, the fact is that I
can use just about any HTTP client to retrieve a document whereas to
get rdf processing clients, agents, whatever, to do it will require
quite a lot of work [1]. This is one reason why, for example, 4store's
arrangement of /sparql/ for read operations and /data/ and /update/
for write operations is *so* much easier to work with than Virtuoso's
OAuth and WebID arrangement - I can just restrict access using all of
the normal tools like apache, nginx, squid, etc..

So in the end we have some work being done to address the perception
that RDF is difficult to work with and on the other hand a suggestion
of widespread putting in place of authentication infrastructure which,
whilst obviously filling a need, stands to make working with the data
behind it more difficult.

How do we balance these two tendencies?

[1] examples of non-WebID aware clients: rapper / rasqal, python
rdflib, curl, the javascript engine in my web browser that doesn't
properly support client certificates, etc.
-- 
William Waitesmailto:w...@styx.org
http://river.styx.org/ww/sip:w...@styx.org
F4B3 39BF E775 CF42 0BAB  3DF0 BE40 A6DF B06F FD45



Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 3:34 PM, Karl Dubost wrote:

Le 21 juin 2011 à 03:49, Martin Hepp a écrit :

Many of the scripts we saw
- ignored robots.txt,
- ignored clear crawling speed limitations in robots.txt,
- did not identify themselves properly in the HTTP request header or lacked 
contact information therein,
- used no mechanisms at all for limiting the default crawling speed and 
re-crawling delays.


Do you have a list of those and how to identify them?
So we can put them in our blocking lists?

.htaccess or Apache config with rules such as:

# added for abusive downloads or not respecting robots.txt
SetEnvIfNoCase User-Agent .*Technorati*. bad_bot
SetEnvIfNoCase User-Agent .*WikioFeedBot*. bad_bot
# [… cut part of my list …]
Order Allow,Deny
Deny from 85.88.12.104
Deny from env=bad_bot
Allow from all





But that doesn't solve the big problem. An Apache module for WebID that 
allows QoS algorithms or heuristics based on Trust Logics is the only 
way this will scale, ultimately. Apache can get with the program, via 
modules. Henry and Joe and a few others are working on keeping Apache in 
step with the new Data Space dimension of the Web :-)


--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Karl Dubost

Le 22 juin 2011 à 10:41, Kingsley Idehen a écrit :
 But that doesn't solve the big problem.

maybe… that solve the resource issue in the meantime ;)
small steps. 

-- 
Karl Dubost - http://dev.opera.com/
Developer Relations  Tools, Opera Software




Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Leigh Dodds
Hi,

On 22 June 2011 15:41, William Waites w...@styx.org wrote:
 What does WebID have to do with JSON? They're somehow representative
 of two competing trends.

 The RDF/JSON, JSON-LD, etc. work is supposed to be about making it
 easier to work with RDF for your average programmer, to remove the
 need for complex parsers, etc. and generally to lower the barriers.

 The WebID arrangement is about raising barriers. Not intended to be
 the same kind of barriers, certainly the intent isn't to make
 programmer's lives more difficult, rather to provide a good way to do
 distributed authentication without falling into the traps of PKI and
 such.

 While I like WebID, and I think it is very elegant, the fact is that I
 can use just about any HTTP client to retrieve a document whereas to
 get rdf processing clients, agents, whatever, to do it will require
 quite a lot of work [1]. This is one reason why, for example, 4store's
 arrangement of /sparql/ for read operations and /data/ and /update/
 for write operations is *so* much easier to work with than Virtuoso's
 OAuth and WebID arrangement - I can just restrict access using all of
 the normal tools like apache, nginx, squid, etc..

 So in the end we have some work being done to address the perception
 that RDF is difficult to work with and on the other hand a suggestion
 of widespread putting in place of authentication infrastructure which,
 whilst obviously filling a need, stands to make working with the data
 behind it more difficult.

 How do we balance these two tendencies?

By recognising that often we just need to use existing technologies
more effectively and more widely, rather than throw more technology at
a problem, thereby creating an even greater education and adoption
problem?

Cheers,

L.

-- 
Leigh Dodds
Programme Manager, Talis Platform
Mobile: 07850 928381
http://kasabi.com
http://talis.com

Talis Systems Ltd
43 Temple Row
Birmingham
B2 5LS



Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Steve Harris
Yes, exactly.

I think that the problem is at least partly (and I say this as an ex-academic) 
that few people in academia have the slightest idea how much it costs to run a 
farm of servers in the Real World™.

From the point of view of the crawler they're trying to get as much data as 
possible in a short a time as possible, but don't realise that the poor guy at 
the other end just got his 95th percentile shot through the roof, and now has a 
several thousand dollar bandwidth bill heading his way.

You can cap bandwidth, but that then might annoy paying customers, which is 
clearly not good.

- Steve

On 2011-06-22, at 12:54, Hugh Glaser wrote:

 Hi Chris.
 One way to do the caching really efficiently:
 http://lists.w3.org/Archives/Public/semantic-web/2007Jun/0012.html
 Which is what rkb has always done.
 But of course caching does not solve the problem of one bad crawler.
 It actually makes it worse.
 You add a cache write cost to the query, without a significant probability of 
 a future cache hit. And increase disk usage.
 
 Hugh
 
 - Reply message -
 From: Christopher Gutteridge c...@ecs.soton.ac.uk
 To: Martin Hepp martin.h...@ebusiness-unibw.org
 Cc: Daniel Herzig her...@kit.edu, semantic-...@w3.org 
 semantic-...@w3.org, public-lod@w3.org public-lod@w3.org
 Subject: Think before you write Semantic Web crawlers
 Date: Wed, Jun 22, 2011 9:18 am
 
 
 
 The difference between these two scenarios is that there's almost no CPU 
 involvement in serving the PDF file, but naive RDF sites use lots of cycles 
 to generate the response to a query for an RDF document.
 
 Right now queries to data.southampton.ac.uk (eg. 
 http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made 
 live, but this is not efficient. My colleague, Dave Challis, has prepared a 
 SPARQL endpoint which caches results which we can turn on if the load gets 
 too high, which should at least mitigate the problem. Very few datasets 
 change in a 24 hours period.
 
 Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are to 
 blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit by 
 DoS-style crazy crawlers. Why should the same not hold if I publish RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF datasets 
 to protect themselves against hungry crawlers and actual DoS attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting 
 Semantic Sitemap information for DoS attacks, focussing on the large dump 
 files.
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your servers 
 and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from site-owners 
 that Semantic Web crawlers from Universities visited their sites in a way 
 close to a denial-of-service attack, i.e., crawling data with maximum 
 bandwidth in a parallelized approach.
 
 It's clear that a single, stupidly written crawler script, run from a 
 powerful University network, can quickly create terrible traffic load.
 
 Many of the scripts we saw
 
 - ignored robots.txt,
 - ignored clear crawling speed limitations in robots.txt,
 - did not identify themselves properly in the HTTP request header or lacked 
 contact information therein,
 - used no mechanisms at all for limiting the default crawling speed and 
 re-crawling delays.
 
 This irresponsible behavior can be the final reason for site-owners to say 
 farewell to academic/W3C-sponsored semantic technology.
 
 So please, please - advise all of your colleagues and students to NOT write 
 simple crawler scripts for the billion triples challenge or whatsoever 
 without familiarizing themselves with the state of the art in friendly 
 crawling.
 
 Best wishes
 
 Martin Hepp
 
 
 
 
 
 
 
 
 --
 Christopher Gutteridge -- http://id.ecs.soton.ac.uk/person/1248
 
 You should read the ECS Web Team blog: http://blogs.ecs.soton.ac.uk/webteam/
 
 

-- 
Steve Harris, CTO, Garlik Limited
1-3 Halford Road, Richmond, TW10 6AW, UK
+44 20 8439 8203  http://www.garlik.com/
Registered in England and Wales 535 7233 VAT # 849 0517 11
Registered office: Thames House, Portsmouth Road, Esher, Surrey, KT10 9AD




Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Kingsley Idehen

On 6/22/11 3:41 PM, William Waites wrote:

While I like WebID, and I think it is very elegant, the fact is that I
can use just about any HTTP client to retrieve a document whereas to
get rdf processing clients, agents, whatever, to do it will require
quite a lot of work [1]. This is one reason why, for example, 4store's
arrangement of/sparql/  for read operations and/data/  and/update/
for write operations is*so*  much easier to work with than Virtuoso's
OAuth and WebID arrangement - I can just restrict access using all of
the normal tools like apache, nginx, squid, etc..

Huh?

WebID and SPARQL is about making an Endpoint with ACLs. ACL membership 
is driven by WebID for people, organizations, or groups (of either).


Don't really want to get into a Virtuoso vs 4-Store argument, but do 
explain to me how the convention you espouse enables me confine access 
to a SPARQL endpoint for:


A person identified by URI based Name (WebID) that a member of a 
foaf:Group (which also has its own WebID).


How does this approach leave ACL membership management to designated 
members of the foaf:Group?


Again, don't wanna do a 4-Store vs Virtuoso, but I really don't get your 
point re. WebID and the fidelity it brings to data access in general. 
Also note, SPARQL endpoints are but one type of data access address. 
WebID protects access to data accessible via Addresses by implicitly 
understanding the difference between a generic Name and a Name 
specifically used a Data Source Address or Location.




--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen







Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 3:44 PM, Karl Dubost wrote:

Le 22 juin 2011 à 10:41, Kingsley Idehen a écrit :

But that doesn't solve the big problem.

maybe… that solve the resource issue in the meantime ;)
small steps.

Yes, but it we make one small viral step. It just scales. Otherwise, we 
will continue to make little steps that don't scale etc..


--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 3:57 PM, Steve Harris wrote:

Yes, exactly.

I think that the problem is at least partly (and I say this as an ex-academic) 
that few people in academia have the slightest idea how much it costs to run a 
farm of servers in the Real World™.

 From the point of view of the crawler they're trying to get as much data as 
possible in a short a time as possible, but don't realise that the poor guy at 
the other end just got his 95th percentile shot through the roof, and now has a 
several thousand dollar bandwidth bill heading his way.

You can cap bandwidth, but that then might annoy paying customers, which is 
clearly not good.


Yes, so we need QoS algorithms or heuristics capable of fine-grained 
partitioning re. Who can do What, When, and Where :-)


Kingsley

- Steve

On 2011-06-22, at 12:54, Hugh Glaser wrote:


Hi Chris.
One way to do the caching really efficiently:
http://lists.w3.org/Archives/Public/semantic-web/2007Jun/0012.html
Which is what rkb has always done.
But of course caching does not solve the problem of one bad crawler.
It actually makes it worse.
You add a cache write cost to the query, without a significant probability of a 
future cache hit. And increase disk usage.

Hugh

- Reply message -
From: Christopher Gutteridgec...@ecs.soton.ac.uk
To: Martin Heppmartin.h...@ebusiness-unibw.org
Cc: Daniel Herzigher...@kit.edu, semantic-...@w3.orgsemantic-...@w3.org, 
public-lod@w3.orgpublic-lod@w3.org
Subject: Think before you write Semantic Web crawlers
Date: Wed, Jun 22, 2011 9:18 am



The difference between these two scenarios is that there's almost no CPU 
involvement in serving the PDF file, but naive RDF sites use lots of cycles to 
generate the response to a query for an RDF document.

Right now queries to data.southampton.ac.uk (eg. 
http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made 
live, but this is not efficient. My colleague, Dave Challis, has prepared a 
SPARQL endpoint which caches results which we can turn on if the load gets too 
high, which should at least mitigate the problem. Very few datasets change in a 
24 hours period.

Martin Hepp wrote:

Hi Daniel,
Thanks for the link! I will relay this to relevant site-owners.

However, I still challenge Andreas' statement that the site-owners are to blame 
for publishing large amounts of data on small servers.

One can publish 10,000 PDF documents on a tiny server without being hit by 
DoS-style crazy crawlers. Why should the same not hold if I publish RDF?

But for sure, it is necessary to advise all publishers of large RDF datasets to 
protect themselves against hungry crawlers and actual DoS attacks.

Imagine if a large site was brought down by a botnet that is exploiting 
Semantic Sitemap information for DoS attacks, focussing on the large dump files.
This could end LOD experiments for that site.


Best

Martin


On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:



Hi Martin,

Have you tried to put a Squid [1]  as reverse proxy in front of your servers 
and use delay pools [2] to catch hungry crawlers?

Cheers,
Daniel

[1] http://www.squid-cache.org/
[2] http://wiki.squid-cache.org/Features/DelayPools

On 21.06.2011, at 09:49, Martin Hepp wrote:



Hi all:

For the third time in a few weeks, we had massive complaints from site-owners 
that Semantic Web crawlers from Universities visited their sites in a way close 
to a denial-of-service attack, i.e., crawling data with maximum bandwidth in a 
parallelized approach.

It's clear that a single, stupidly written crawler script, run from a powerful 
University network, can quickly create terrible traffic load.

Many of the scripts we saw

- ignored robots.txt,
- ignored clear crawling speed limitations in robots.txt,
- did not identify themselves properly in the HTTP request header or lacked 
contact information therein,
- used no mechanisms at all for limiting the default crawling speed and 
re-crawling delays.

This irresponsible behavior can be the final reason for site-owners to say 
farewell to academic/W3C-sponsored semantic technology.

So please, please - advise all of your colleagues and students to NOT write simple 
crawler scripts for the billion triples challenge or whatsoever without familiarizing 
themselves with the state of the art in friendly crawling.

Best wishes

Martin Hepp








--
Christopher Gutteridge -- http://id.ecs.soton.ac.uk/person/1248

You should read the ECS Web Team blog: http://blogs.ecs.soton.ac.uk/webteam/





--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Dave Reynolds
On Wed, 2011-06-22 at 15:52 +0100, Leigh Dodds wrote: 
 Hi,
 
 On 22 June 2011 15:41, William Waites w...@styx.org wrote:
  What does WebID have to do with JSON? They're somehow representative
  of two competing trends.
 
  The RDF/JSON, JSON-LD, etc. work is supposed to be about making it
  easier to work with RDF for your average programmer, to remove the
  need for complex parsers, etc. and generally to lower the barriers.
 
  The WebID arrangement is about raising barriers. Not intended to be
  the same kind of barriers, certainly the intent isn't to make
  programmer's lives more difficult, rather to provide a good way to do
  distributed authentication without falling into the traps of PKI and
  such.
 
  While I like WebID, and I think it is very elegant, the fact is that I
  can use just about any HTTP client to retrieve a document whereas to
  get rdf processing clients, agents, whatever, to do it will require
  quite a lot of work [1]. This is one reason why, for example, 4store's
  arrangement of /sparql/ for read operations and /data/ and /update/
  for write operations is *so* much easier to work with than Virtuoso's
  OAuth and WebID arrangement - I can just restrict access using all of
  the normal tools like apache, nginx, squid, etc..
 
  So in the end we have some work being done to address the perception
  that RDF is difficult to work with and on the other hand a suggestion
  of widespread putting in place of authentication infrastructure which,
  whilst obviously filling a need, stands to make working with the data
  behind it more difficult.
 
  How do we balance these two tendencies?
 
 By recognising that often we just need to use existing technologies
 more effectively and more widely, rather than throw more technology at
 a problem, thereby creating an even greater education and adoption
 problem?

+1

Don't raise barriers to linked data use/publication by tying it to
widespread adoption and support for WebID.

Dave





Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Kingsley Idehen

On 6/22/11 4:08 PM, Dave Reynolds wrote:

On Wed, 2011-06-22 at 15:52 +0100, Leigh Dodds wrote:

Hi,

On 22 June 2011 15:41, William Waitesw...@styx.org  wrote:

What does WebID have to do with JSON? They're somehow representative
of two competing trends.

The RDF/JSON, JSON-LD, etc. work is supposed to be about making it
easier to work with RDF for your average programmer, to remove the
need for complex parsers, etc. and generally to lower the barriers.

The WebID arrangement is about raising barriers. Not intended to be
the same kind of barriers, certainly the intent isn't to make
programmer's lives more difficult, rather to provide a good way to do
distributed authentication without falling into the traps of PKI and
such.

While I like WebID, and I think it is very elegant, the fact is that I
can use just about any HTTP client to retrieve a document whereas to
get rdf processing clients, agents, whatever, to do it will require
quite a lot of work [1]. This is one reason why, for example, 4store's
arrangement of /sparql/ for read operations and /data/ and /update/
for write operations is *so* much easier to work with than Virtuoso's
OAuth and WebID arrangement - I can just restrict access using all of
the normal tools like apache, nginx, squid, etc..

So in the end we have some work being done to address the perception
that RDF is difficult to work with and on the other hand a suggestion
of widespread putting in place of authentication infrastructure which,
whilst obviously filling a need, stands to make working with the data
behind it more difficult.

How do we balance these two tendencies?

By recognising that often we just need to use existing technologies
more effectively and more widely, rather than throw more technology at
a problem, thereby creating an even greater education and adoption
problem?

+1

Don't raise barriers to linked data use/publication by tying it to
widespread adoption and support for WebID.


-1

You are misunderstanding WebID and what it delivers.

I am popping out, but I expect a response. Should Henry not put this 
misconception to REST, I'll certainly reply.


Got to go do some walking for now :-)

Dave







--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread William Waites
* [2011-06-22 16:00:49 +0100] Kingsley Idehen kide...@openlinksw.com écrit:

] explain to me how the convention you espouse enables me confine access 
] to a SPARQL endpoint for:
] 
] A person identified by URI based Name (WebID) that a member of a 
] foaf:Group (which also has its own WebID).

This is not a use case I encounter much. Usually I have some
application code that needs write access to the store and some public
code (maybe javascript in a browser, maybe some program run by a third
party) that needs read access.

If the answer is to teach my application code about WebID, it's going
to be a hard sell because really I want to be working on other things
than protocol plumbing.

If you then go further and say that *all* access to the endpoint needs
to use WebID because of resource-management issues, then every client
now needs to do a bunch of things that end with shaving a yak before
they can even start on working on whatever they were meant to be
working on.

On the other hand, arranging things so that access control can be done
by existing tools without burdening the clients is a lot easier, if
less general. And easier is what we want working with RDF to be.

Cheers,
-w

-- 
William Waitesmailto:w...@styx.org
http://river.styx.org/ww/sip:w...@styx.org
F4B3 39BF E775 CF42 0BAB  3DF0 BE40 A6DF B06F FD45



WebID and client tools - was: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Henry Story

On 22 Jun 2011, at 16:41, William Waites wrote:

 
 [1] examples of non-WebID aware clients: rapper / rasqal, python
 rdflib, curl, the javascript engine in my web browser that doesn't
 properly support client certificates, etc.

curl is WebID aware. You just need to get yourself a certificate for your 
crawler, and then use

  -E/--cert  certificate[:password] 

arguments to pass that certificate if the server requests it. 

The specs for HTTPS client certs are so old and well established that it is 
built by default into most libraries. So it would not take a lot to expose it, 
if it is not already in all the libs you mention.

But thanks for this new FAQ [1]. We'll try to fill in the details on how to 
work with the libs above using webid.

There is a Javascript layer for https too, but what is the point of doing that 
there? Let the browser do the https for you.

Henry

[1] http://www.w3.org/wiki/Foaf%2Bssl/FAQ

Social Web Architect
http://bblfish.net/




Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Henry Story

On 22 Jun 2011, at 17:14, William Waites wrote:

 * [2011-06-22 16:00:49 +0100] Kingsley Idehen kide...@openlinksw.com écrit:
 
 ] explain to me how the convention you espouse enables me confine access 
 ] to a SPARQL endpoint for:
 ] 
 ] A person identified by URI based Name (WebID) that a member of a 
 ] foaf:Group (which also has its own WebID).
 
 This is not a use case I encounter much. Usually I have some
 application code that needs write access to the store and some public
 code (maybe javascript in a browser, maybe some program run by a third
 party) that needs read access.
 
 If the answer is to teach my application code about WebID, it's going
 to be a hard sell because really I want to be working on other things
 than protocol plumbing.

So you're in luck. https is shipped in all client libraries, so you just need
to get your application a webid certificate. That should be as easy as one post 
request
to get it. At least for browsers it's a one click affair for the end user, as 
shown
here

   http://bblfish.net/blog/2011/05/25/

It would be easy to do the same for robots. In fact that is why at the 
University
of Manchester Bruno Harbulot and Mike Jones are using WebID for their Grid 
computing 
work, because it makes access control to the grid so much easier that any of 
the other
top heavy technologies available.

 If you then go further and say that *all* access to the endpoint needs
 to use WebID because of resource-management issues, then every client
 now needs to do a bunch of things that end with shaving a yak before
 they can even start on working on whatever they were meant to be
 working on.

You can be very flexible there.  If users have WebId you give them a better 
service.
Seems fair deal. It can also be very flexible. You don't need all your site to 
be WebID
enabled. You could use cookie auth on http endpoints, and for clients that 
don't have
a cookie redirect them to an https endpoint where they can auth with webid. If 
they don't
ask them to auth with somithing like OpenId. I'd say pretty soon your crawlers 
and users will
be a lot happier with WebID.

 On the other hand, arranging things so that access control can be done
 by existing tools without burdening the clients is a lot easier, if
 less general. And easier is what we want working with RDF to be.

All your tools probably already are webId enabled. It's just a matter now of
giving a foaf profile to yourself and robots, getting a cert with the webid 
inthere,
and getting going. Seems to be that that's a lot easier than building crawlers, 
or semweb clients, or semweb servers, or pretty much anything.

Henry

 
 Cheers,
 -w
 
 -- 
 William Waitesmailto:w...@styx.org
 http://river.styx.org/ww/sip:w...@styx.org
 F4B3 39BF E775 CF42 0BAB  3DF0 BE40 A6DF B06F FD45
 

Social Web Architect
http://bblfish.net/




CfP: ACM RecSys 2011 International Workshop on Information Heterogeneity and Fusion in Recommender Systems (HetRec 2011)

2011-06-22 Thread Iván Cantador
[Apologies if you receive this more than once]


2nd Call for Papers

2nd International Workshop on Information Heterogeneity and Fusion in
Recommender Systems (HetRec 2011)
27th October 2011 | Chicago, IL, USA 
http://ir.ii.uam.es/hetrec2011 

Held in conjunction with the
5th ACM Conference on Recommender Systems (RecSys 2011) 
http://recsys.acm.org/2011 

 

+++
Important dates
+++

* Paper submission: 25 July 2011
* Notification of acceptance:   19 August 2011
* Camera-ready version due: 12 September 2011
* HetRec 2011 Workshop: 27 October 2011

++
Motivation
++

In recent years, increasing attention has been given to finding ways for
combining, integrating and mediating heterogeneous sources of information
for the purpose of providing better personalized services in many
information seeking and e-commerce applications. Information heterogeneity
can indeed be identified in any of the pillars of a recommender system: the
modeling of user preferences, the description of resource contents, the
modeling and exploitation of the context in which recommendations are made,
and the characteristics of the suggested resource lists.

Almost all current recommender systems are designed for specific domains and
applications, and thus usually try to make best use of a local user model,
using a single kind of personal data, and without explicitly addressing the
heterogeneity of the existing personal information that may be freely
available (on social networks, homepages, etc.). Recognizing this
limitation, among other issues: a) user models could be based on different
types of explicit and implicit personal preferences, such as ratings, tags,
textual reviews, records of views, queries, and purchases; b) recommended
resources may belong to several domains and media, and may be described with
multilingual metadata; c) context could be modeled and exploited in
multi-dimensional feature spaces; d) and ranked recommendation lists could
be diverse according to particular user preferences and resource attributes,
oriented to groups of users, and driven by multiple user evaluation
criteria.

The aim of HetRec workshop is to bring together students, faculty,
researchers and professionals from both academia and industry who are
interested in addressing any of the above forms of information heterogeneity
and fusion in recommender systems.

The workshop goals are broad. We would like to raise awareness of the
potential of using multiple sources of information, and look for sharing
expertise and suitable models and techniques. Another dire need is for
strong datasets, and one of our aims is to establish benchmarks and standard
datasets on which the problems could be investigated.

++
Topics of interest
++

The goal of the workshop is to bring together researchers and practitioners
interested in addressing the challenges posed by information heterogeneity
in recommender systems, and studying information fusion in this context.

Topics of interest include, but are not limited to:

* Fusion of user profiles from different representations, e.g. ratings,
text reviews, tags, and bookmarks
* Combination of short- and long-term user preferences
* Combination of different types of user preferences: tastes, interests,
needs, goals, mood
* Cross-domain recommendations, based on user preferences about
different interest aspects, e.g. by merging movie and music tastes
* Cross-representation recommendations, considering diverse sources of
user preferences: explicit and implicit feedback
* Recommendation of resources of different nature: news, reviews,
scientific papers, etc.
* Recommendation of resources belonging to different multimedia: text,
image, audio, video
* Recommendation of diverse resources, e.g. according to content
attributes, and user consuming behaviors
* Recommendation of resources annotated in different languages
* Contextualization of multiple user preferences, e.g. by distinguishing
user preferences at work and on holidays
* Cross-context recommendations, e.g. by merging information about
location, time and social aspects
* Multi-dimensional recommendation based on several contextual features,
e.g. physical and social environment, device and network settings, and
external events
* Multi-criteria recommendation, exploiting ratings and evaluations
about multiple user/item characteristics
* Group recommendation, oriented to several users, e.g. suggesting
tourist attractions to a group of friends, and suggesting a TV show to a
family

+++
Keynote
+++

We are pleased to announce that Yehuda Koren, from Yahoo! Research, will be
our invited speaker.

The title of his talk is I Want to Answer, Who Has a Question? Yahoo!
Answers Recommender 

Re: WebID vs. JSON (Was: Re: Think before you write Semantic Web crawlers)

2011-06-22 Thread Kingsley Idehen

On 6/22/11 4:14 PM, William Waites wrote:

* [2011-06-22 16:00:49 +0100] Kingsley Idehenkide...@openlinksw.com  écrit:

] explain to me how the convention you espouse enables me confine access
] to a SPARQL endpoint for:
]
] A person identified by URI based Name (WebID) that a member of a
] foaf:Group (which also has its own WebID).

This is not a use case I encounter much. Usually I have some
application code that needs write access to the store and some public
code (maybe javascript in a browser, maybe some program run by a third
party) that needs read access.


I am assuming you seek multiple users of your end product (the 
application), right?


I assume all users aren't equal i.e., they have varying profiles, right?

If the answer is to teach my application code about WebID, it's going
to be a hard sell because really I want to be working on other things
than protocol plumbing.


Remember, I like to take the problem solving approach to technology. 
Never technology for the sake of it, never.


There is a fundamental problem: you seek 1 users of your apps. All 
users aren't the same, profile wise.


Simple case in point, right here, and right now. This thread is about a 
critical challenge (always there btw) that Linked Data propagation 
unveils. The very same problems hit us in the early '90s re. ODBC i.e., 
how do we control access to data bearing in mind ODBC application user 
profile variations. Should anyone be able to access pensions and payroll 
data to make a very obvious example.


The gaping security hole that ODBC introduced to the enterprise is still 
doing damage to this very day. I won't mention names, but as you hear 
about security breaches, do some little digging about what's behind many 
of these systems. Hint: a relational database, and free ODBC, JDBC, 
OLE-DB, ADO.NET providers, in many cases. Have one of those libraries  
on a system, you can get into the RDBMS via social engineering (in the 
absolute worst case, or throttle with CPUs for passwords).


Way back then we use Windows INI structure to construct a graph based 
data representation format that we called session rules book. Via 
these rules we enabled organizations to say: Kingsley can only access 
records in certain ODBC/JDBC/OLE-DB/ADO.NET accessible databases if he 
met certain criteria that included the IP address he logs in from, his 
username, client application name, arbitrary identifiers that the system 
owner could conjure up etc.. The only drag for us what it was little 
OpenLink rather than a behemoth like Microsoft.


When we encountered RDF and the whole Semantic Web vision we realized 
there was a standardized route for addressing these fundamental issues. 
This is why WebID is simply a major deal. It is inherently quite 
contradictory to push Linked Data and push-back at WebID. That's only 
second to rejecting essence of URI abstraction by conflating Names and 
Addresses re. fine grained data access that address troubling problems 
of yore.




If you then go further and say that *all* access to the endpoint needs
to use WebID because of resource-management issues, then every client
now needs to do a bunch of things that end with shaving a yak before
they can even start on working on whatever they were meant to be
working on.



No.

This is what we (WebID implementers) are saying:

1. Publish Linked Data
2. Apply Linked Data prowess to the critical issue of controlled access 
to Linked Data Spaces.


Use Linked Data to solve a real problem. In doing so we'll achieve the 
critical mass we all seek because the early adopters of Linked Data will 
be associated with:


1. Showing how Linked Data solves a real problem
2. Using Linked Data to make its use and consumption easier for others 
who seek justification and use case examples en route to full investment.



On the other hand, arranging things so that access control can be done
by existing tools without burdening the clients is a lot easier, if
less general. And easier is what we want working with RDF to be.


It has nothing to do with RDF. It has everything to do with Linked Data 
i.e., Data Objects endowed with Names that resolve to their 
Representations. Said representations take the form of EAV/SPO based 
graphs. RDF is one of the options for achieving this goal via a syntax 
with high semantic fidelity (most of that comes from granularity 
covering datatypes and locale issues).


What people want, and have always sought is: open access to relevant 
data from platforms and tools of their choice without any performance or 
security compromises. HTTP, URIs, and exploitation of full URI 
abstraction as mechanism for graph based whole data representation, 
without graph format/syntax distractions is the beachhead we need right 
now. The semantic fidelity benefits of RDF re. datatypes and locale 
issues comes after that. Thus, first goal is to actually simplify Linked 
Data, and make its use and exploitation practical,  starting with 
appreciation of 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Martin Hepp
Glenn:

 If there isn't, why not? We're the Semantic Web, dammit. If we aren't the 
 masters of data interoperability, what are we?
The main question is: Is the Semantic Web an evolutionary improvement of the 
Web, the Web understood as an ecosystem comprising protocols, data models, 
people, and economics - or is it a tiny special interest branch.

As said: I bet a bottle of champagne that the academic Semantic Web community's 
technical proposals will never gain more than 10 % market share among real 
site-owners, because of
- unnecessary complexity (think of the simplicity of publishing an HTML page 
vs. following LOD publishing principles),
- bad design decisions (e.g explicit datatyping of data instances in RDFa),
- poor documentation for non-geeks, and
- a lack of understanding of the economics of technology diffusion.

Never ever.

Best

Martin

On Jun 22, 2011, at 3:18 PM, glenn mcdonald wrote:

 From my perspective as the designer of a system that both consumes and 
 publishes data, the load/burden issue here is not at all particular to the 
 semantic web. Needle obeys robots.txt rules, but that's a small deal 
 compared to the difficulty of extracting whole data from sites set up to 
 deliver it only in tiny pieces. I'd say about 98% of the time I can describe 
 the data I want from a site with a single conceptual query. Indeed, once 
 I've got the data into Needle I can almost always actually produce that 
 query. But on the source site, I usually can't, and thus we are forced to 
 waste everybody's time navigating the machines through superfluous 
 presentation rendering designed for people. 10-at-a-time results lists, 
 interminable AJAX refreshes, animated DIV reveals, grafting back together 
 the splintered bits of tree-traversals, etc. This is all absurdly 
 unnecessary. Why is anybody having to crawl an open semantic-web dataset? 
 Isn't there a download link, and/or a SPARQL endpoint? If there isn't, why 
 not? We're the Semantic Web, dammit. If we aren't the masters of data 
 interoperability, what are we?
 
 glenn
 (www.needlebase.com)




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Martin Hepp
Hi Andreas:

Please make a survey among typical Web site owners on how many of them have

1. access to this level of server configuration and
2. the skills necessary to implement these recommendations.

The WWW was anti-pedantic by design. This was the root of its success. The 
pedants were the traditional SGML/Hypertext communities. Why are we breeding 
new pedants?

Martin

On Jun 22, 2011, at 11:44 AM, Andreas Harth wrote:

 Hi Christopher,
 
 On 06/22/2011 10:14 AM, Christopher Gutteridge wrote:
 Right now queries to data.southampton.ac.uk (eg.
 http://data.southampton.ac.uk/products-and-services/CupCake.rdf ) are made 
 live,
 but this is not efficient. My colleague, Dave Challis, has prepared a SPARQL
 endpoint which caches results which we can turn on if the load gets too high,
 which should at least mitigate the problem. Very few datasets change in a 24
 hours period.
 
 setting the Expires header and enabling mod_cache in Apache httpd (or adding
 a Squid proxy in front of the HTTP server) works quite well in these cases.
 
 Best regards,
 Andreas.
 




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Andreas Harth

Hi Martin,

On 06/22/2011 09:08 PM, Martin Hepp wrote:

Please make a survey among typical Web site owners on how many of them have

1. access to this level of server configuration and

 2. the skills necessary to implement these recommendations.

d'accord .

But the case we're discussing there's also:

3. publishes millions of pages

I am glad you brought up the issue, as there are several data providers
out there (some with quite prominent names) with hundreds of millions of
triples, but unable to sustain lookups every couple of seconds or so.

I am very much in favour of amateur web enthusiasts (I would like to claim
I've started as one).  Unfortunately, you get them on both ends, publishers
and consumers.  Postel's law applies to both, I guess.

Best regards,
Andreas.



Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Kingsley Idehen

On 6/22/11 8:29 PM, Andreas Harth wrote:

I am glad you brought up the issue, as there are several data providers
out there (some with quite prominent names) with hundreds of millions of
triples, but unable to sustain lookups every couple of seconds or so. 

But that's quite general a statement you make.

What about the fact that the data providers have configured their 
systems such that lookups aren't performed by anyone every couple of 
seconds or so? That pinhole is enough fooder for DoS in Web context :-)


--

Regards,

Kingsley Idehen 
President  CEO
OpenLink Software
Web: http://www.openlinksw.com
Weblog: http://www.openlinksw.com/blog/~kidehen
Twitter/Identi.ca: kidehen








Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Sebastian Schaffert
Martin,

I followed the thread a bit, and I have just a small and maybe naive question: 
what use is a Linked Data Web that does not even scale to the access of 
crawlers? And how to we expect agents to use Linked Data if we cannot provide 
technology that scales?

Your complaint sounds to me a bit like help, too many clients access my data. 
I think worse things could happen. What we need to do is improve our 
technology, and not whining about people trying to use our data. Even though it 
is not good if people stop *providing* Linked Data, it is also not good if 
people stop *using* Linked Data. And I find your approach of stopping to send 
pings counter-productive.

My 2 cents to the discussion ... :-)

Greetings,

Sebastian


Am 22.06.2011 um 20:57 schrieb Martin Hepp:

 Jiri:
 The crawlers causing problems were run by Universities, mostly in the context 
 of ISWC submissions. No need to cast any doubt on that.
 
 All:
 As a consequence to those events, I will not publish sitemaps etc. of future 
 GoodRelations datasets on these lists, but just inform non-toy consumers.
 If you consider yourself a non-toy consumer of e-commerce data, please send 
 me an e-mail, and we will add you to out ping chain.
 
 We will also stop sending pings to PTSW, Watson, Swoogle, et al., because 
 they will just expose sites adopting GoodRelations and related technology to 
 academic crawling.
 
 In the meantime, I recommend the LOD bubble diagram sources for 
 self-referential research.
 
 Best
 M. Hepp
 
 
 
 On Jun 22, 2011, at 4:03 PM, Jiří Procházka wrote:
 
 I understand that, but I doubt your conclusion, that those crawlers are
 targeting semantic web, since like you said they don't even properly
 identify themselves and as far as I know, on Universities also regular
 web search and crawling is researched. Maybe lot of them are targeting
 semantic web, but we should look at all measures to conserve bandwidth,
 from avoiding regular web crawler interest, aiding infrastructure like
 Ping the Semantic Web to optimizing delivery and even distribution of
 the data among resouces.
 
 Best,
 Jiri
 
 On 06/22/2011 03:21 PM, Martin Hepp wrote:
 Thanks, Jiri, but the load comes from academic crawler prototypes firing 
 from broad University infrastructures.
 Best
 Martin
 
 
 On Jun 22, 2011, at 12:40 PM, Jiří Procházka wrote:
 
 I wonder, are ways to link RDF data so that convential crawlers do not
 crawl it, but only the semantic web aware ones do?
 I am not sure how the current practice of linking by link tag in the
 html headers could cause this, but it may be case that those heavy loads
 come from a crawlers having nothing to do with semantic web...
 Maybe we should start linking to our rdf/xml, turtle, ntriples files and
 publishing sitemap info in RDFa...
 
 Best,
 Jiri
 
 On 06/22/2011 09:00 AM, Steve Harris wrote:
 While I don't agree with Andreas exactly that it's the site owners fault, 
 this is something that publishers of non-semantic data have to deal with.
 
 If you publish a large collection of interlinked data which looks 
 interesting to conventional crawlers and is expensive to generate, 
 conventional web crawlers will be all over it. The main difference is 
 that a greater percentage of those are written properly, to follow 
 robots.txt and the guidelines about hit frequency (maximum 1 request per 
 second per domain, no parallel crawling).
 
 Has someone published similar guidelines for semantic web crawlers?
 
 The ones that don't behave themselves get banned, either in robots.txt, 
 or explicitly by the server. 
 
 - Steve
 
 On 2011-06-22, at 06:07, Martin Hepp wrote:
 
 Hi Daniel,
 Thanks for the link! I will relay this to relevant site-owners.
 
 However, I still challenge Andreas' statement that the site-owners are 
 to blame for publishing large amounts of data on small servers.
 
 One can publish 10,000 PDF documents on a tiny server without being hit 
 by DoS-style crazy crawlers. Why should the same not hold if I publish 
 RDF?
 
 But for sure, it is necessary to advise all publishers of large RDF 
 datasets to protect themselves against hungry crawlers and actual DoS 
 attacks.
 
 Imagine if a large site was brought down by a botnet that is exploiting 
 Semantic Sitemap information for DoS attacks, focussing on the large 
 dump files. 
 This could end LOD experiments for that site.
 
 
 Best
 
 Martin
 
 
 On Jun 21, 2011, at 10:24 AM, Daniel Herzig wrote:
 
 
 Hi Martin,
 
 Have you tried to put a Squid [1]  as reverse proxy in front of your 
 servers and use delay pools [2] to catch hungry crawlers?
 
 Cheers,
 Daniel
 
 [1] http://www.squid-cache.org/
 [2] http://wiki.squid-cache.org/Features/DelayPools
 
 On 21.06.2011, at 09:49, Martin Hepp wrote:
 
 Hi all:
 
 For the third time in a few weeks, we had massive complaints from 
 site-owners that Semantic Web crawlers from Universities visited their 
 sites in a way close to a denial-of-service attack, i.e., crawling 
 data with maximum 

Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Dave Challis

On 22/06/11 16:05, Kingsley Idehen wrote:

On 6/22/11 3:57 PM, Steve Harris wrote:

Yes, exactly.

I think that the problem is at least partly (and I say this as an
ex-academic) that few people in academia have the slightest idea how
much it costs to run a farm of servers in the Real World™.

From the point of view of the crawler they're trying to get as much
data as possible in a short a time as possible, but don't realise that
the poor guy at the other end just got his 95th percentile shot
through the roof, and now has a several thousand dollar bandwidth bill
heading his way.

You can cap bandwidth, but that then might annoy paying customers,
which is clearly not good.


Yes, so we need QoS algorithms or heuristics capable of fine-grained
partitioning re. Who can do What, When, and Where :-)

Kingsley


There are plenty of these around when it comes to web traffic in 
general.  For apache, I can think of ModSecurity 
(http://www.modsecurity.org/) and mod_evasive 
(http://www.zdziarski.com/blog/?page_id=442).


Both of these will look at traffic patterns and dynamically blacklist as 
needed.


ModSecurity also allows for custom rules to be written depending on 
get/post content, so it should be perfectly feasible to set up rules 
based on estimated/actual query cost (e.g. blacklist if client makes  X 
requests per Y mins which return  Z triples).


Can't see any reason why a hybrid approach couldn't be used, e.g. apply 
rules to unauthenticated traffic, and auto-whitelist clients identifying 
themselves via WebID.


--
Dave Challis
d...@ecs.soton.ac.uk




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Lin Clark
On Wed, Jun 22, 2011 at 9:33 PM, Sebastian Schaffert 
sebastian.schaff...@salzburgresearch.at wrote:


 Your complaint sounds to me a bit like help, too many clients access my
 data.


I'm sure that Martin is really tired of saying this, so I will reiterate for
him: It wasn't his data, they weren't his servers. He's speaking on behalf
of people who aren't part of our insular community... people who don't have
a compelling reason to subsidize a PhD student's Best Paper award with their
own dollars and bandwidth.

Agents can use Linked Data just fine without firing 150 requests per second
at a server. There are TONs of use cases that do not require that kind of
server load.

-- 
Lin Clark
DERI, NUI Galway http://www.deri.ie/

lin-clark.com
twitter.com/linclark


Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Alexandre Passant

On 22 Jun 2011, at 22:49, Richard Cyganiak wrote:

 On 21 Jun 2011, at 10:44, Martin Hepp wrote:
 PS: I will not release the IP ranges from which the trouble originated, but 
 rest assured, there were top research institutions among them.
 
 The right answer is: name and shame. That is the way to teach them.

You may have find the right word: teach.
We've (as academic) given tutorials on how to publish and consume LOD, lots of 
things about best practices for publishing, but not much about consuming.
Why not simply coming with reasonable guidelines for this, that should also be 
taught in institutes / universities where people use LOD, and in tutorials 
given in various conferences.

m2c

Alex.

 
 Like Karl said, we should collect information about abusive crawlers so that 
 site operators can defend themselves. It won't be *that* hard to research and 
 collect the IP ranges of offending universities.
 
 I started a list here:
 http://www.w3.org/wiki/Bad_Crawlers
 
 The list is currently empty. I hope it stays that way.
 
 Thank you all,
 Richard

--
Dr. Alexandre Passant, 
Social Software Unit Leader
Digital Enterprise Research Institute, 
National University of Ireland, Galway






Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Henry Story

On 23 Jun 2011, at 00:11, Alexandre Passant wrote:

 
 On 22 Jun 2011, at 22:49, Richard Cyganiak wrote:
 
 On 21 Jun 2011, at 10:44, Martin Hepp wrote:
 PS: I will not release the IP ranges from which the trouble originated, but 
 rest assured, there were top research institutions among them.
 
 The right answer is: name and shame. That is the way to teach them.
 
 You may have find the right word: teach.
 We've (as academic) given tutorials on how to publish and consume LOD, lots 
 of things about best practices for publishing, but not much about consuming.
 Why not simply coming with reasonable guidelines for this, that should also 
 be taught in institutes / universities where people use LOD, and in tutorials 
 given in various conferences.

That is of course a good idea. But longer term you don't want to teach that 
way. It's too time consuming. You need the machines to do the teaching. 

Think about Facebook. How did 500 million people go to use it? Because they 
were introduced by friends, by using it, but not by doing tutorials and going 
to courses. The system itself teaches people how to use it. 

So the same way, if you want to teach people linked data, get the social web 
going and they will learn the rest by themselves. If you want to teach crawlers 
to behave, make bad behaviour uninteresting. Create a game and rules where good 
behaviour are rewarded and bad behaviour has the opposite effect.

This is why I think using WebID can help. You can use the information to build 
lists and rankings of good and bad crawlers, people with good crawlers get to 
present papers and crawling confs, bad crawlers get throttled out of crawling.  
Make it so that the system can grow beyond academic and teaching settings, into 
the world of billions of users spread across the world, living in different 
political institutions and speaking different languages. We have had good 
crawling practices since the beginning of the web, but you need to make them 
evident and self teaching.

EG. A crawler that crawls to much will get slowed down, and redirected to pages 
on crawling behavior, written and translated into every single language on the 
planet.

Henry


 
 m2c
 
 Alex.
 
 
 Like Karl said, we should collect information about abusive crawlers so that 
 site operators can defend themselves. It won't be *that* hard to research 
 and collect the IP ranges of offending universities.
 
 I started a list here:
 http://www.w3.org/wiki/Bad_Crawlers
 
 The list is currently empty. I hope it stays that way.
 
 Thank you all,
 Richard
 
 --
 Dr. Alexandre Passant, 
 Social Software Unit Leader
 Digital Enterprise Research Institute, 
 National University of Ireland, Galway
 
 
 
 

Social Web Architect
http://bblfish.net/




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Karl Dubost

Le 22 juin 2011 à 18:11, Alexandre Passant a écrit :
 Why not simply coming with reasonable guidelines

started
http://www.w3.org/wiki/Write_Web_Crawler

-- 
Karl Dubost - http://dev.opera.com/
Developer Relations  Tools, Opera Software




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Sebastian Schaffert

Am 22.06.2011 um 23:01 schrieb Lin Clark:

 On Wed, Jun 22, 2011 at 9:33 PM, Sebastian Schaffert 
 sebastian.schaff...@salzburgresearch.at wrote:
 
 Your complaint sounds to me a bit like help, too many clients access my 
 data.
 
 I'm sure that Martin is really tired of saying this, so I will reiterate for 
 him: It wasn't his data, they weren't his servers. He's speaking on behalf of 
 people who aren't part of our insular community... people who don't have a 
 compelling reason to subsidize a PhD student's Best Paper award with their 
 own dollars and bandwidth.

And what about those companies subsidizing PhD students who write crawlers for 
the normal Web? Like Larry Page in 1998?

 
 Agents can use Linked Data just fine without firing 150 requests per second 
 at a server. There are TONs of use cases that do not require that kind of 
 server load.

And what if in the future 100.000 software agents will access servers? We will 
have the scalability issue eventually even without crawlers, so let's try to 
solve it. In the eyeball web, there are also crawlers without too much of a 
problem, and if Linked Data is to be successful we need to do the same.

Greetings,

Sebastian
-- 
| Dr. Sebastian Schaffert  sebastian.schaff...@salzburgresearch.at
| Salzburg Research Forschungsgesellschaft  http://www.salzburgresearch.at
| Head of Knowledge and Media Technologies Group  +43 662 2288 423
| Jakob-Haringer Strasse 5/II
| A-5020 Salzburg




Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Juan Sequeda
On Wed, Jun 22, 2011 at 7:08 PM, Sebastian Schaffert 
sebastian.schaff...@salzburgresearch.at wrote:


 Am 22.06.2011 um 23:01 schrieb Lin Clark:

  On Wed, Jun 22, 2011 at 9:33 PM, Sebastian Schaffert 
 sebastian.schaff...@salzburgresearch.at wrote:
 
  Your complaint sounds to me a bit like help, too many clients access my
 data.
 
  I'm sure that Martin is really tired of saying this, so I will reiterate
 for him: It wasn't his data, they weren't his servers. He's speaking on
 behalf of people who aren't part of our insular community... people who
 don't have a compelling reason to subsidize a PhD student's Best Paper award
 with their own dollars and bandwidth.

 And what about those companies subsidizing PhD students who write crawlers
 for the normal Web? Like Larry Page in 1998?


Talking to some friends at Stanford, they told me some of the problems that
they would go through initially. For example, many website owners would
directly call Stanford and threaten to sue them because of google's
crawlers.




 
  Agents can use Linked Data just fine without firing 150 requests per
 second at a server. There are TONs of use cases that do not require that
 kind of server load.

 And what if in the future 100.000 software agents will access servers? We
 will have the scalability issue eventually even without crawlers, so let's
 try to solve it. In the eyeball web, there are also crawlers without too
 much of a problem, and if Linked Data is to be successful we need to do the
 same.

 Greetings,

 Sebastian
 --
 | Dr. Sebastian Schaffert  sebastian.schaff...@salzburgresearch.at
 | Salzburg Research Forschungsgesellschaft  http://www.salzburgresearch.at
 | Head of Knowledge and Media Technologies Group  +43 662 2288 423
 | Jakob-Haringer Strasse 5/II
 | A-5020 Salzburg





Re: Think before you write Semantic Web crawlers

2011-06-22 Thread Juan Sequeda


 You may have find the right word: teach.
 We've (as academic) given tutorials on how to publish and consume LOD, lots
 of things about best practices for publishing, but not much about consuming.
 Why not simply coming with reasonable guidelines for this, that should also
 be taught in institutes / universities where people use LOD, and in
 tutorials given in various conferences.



Need to put my publicity hat on:

Submit papers to the Consuming Linked Data Workshop!

http://km.aifb.kit.edu/ws/cold2011/

:)




 m2c

 Alex.

 
  Like Karl said, we should collect information about abusive crawlers so
 that site operators can defend themselves. It won't be *that* hard to
 research and collect the IP ranges of offending universities.
 
  I started a list here:
  http://www.w3.org/wiki/Bad_Crawlers
 
  The list is currently empty. I hope it stays that way.
 
  Thank you all,
  Richard

 --
 Dr. Alexandre Passant,
 Social Software Unit Leader
 Digital Enterprise Research Institute,
 National University of Ireland, Galway







Reminder: CfP Special Issue SEMANTIC MULTIMEDIA -- International Journal of Semantic Computing (IJSC)

2011-06-22 Thread Harald Sack

Apologies for any cross postings

--

Reminder: Call for Papers - International Journal of Semantic Computing (IJSC)

Special Issue on SEMANTIC MULTIMEDIA

http://www.worldscinet.com/ijsc/mkt/callforpapers.shtml

--

SCOPE

In the new millennium Multimedia Computing plays an increasingly important role
as more and more users produce and share a constantly growing amount of 
multimedia documents. The sheer number of documents available in large media 
repositories or even the World Wide Web makes indexing and retrieval of 
multimedia documents as well as browsing and annotation more important tasks 
than ever before. Research in this area is of great importance because of the 
very limited understanding of the semantics of such data sources as well as the
limited ways in which they can be accessed by the users today. The field of
Semantic Computing has much to offer with respect to these challenges. This 
special issue invites articles that bring together Semantic Computing and 
Multimedia to address the challenges arising by the constant growth of 
Multimedia.



AREAS OF INTEREST INCLUDE (but are not limited to):

Semantics in Multimedia

* The Role of Multimedia Objects in the Semantic Web
* Multimedia Ontologies and Infrastructures
* Integration of Multimedia Processing and Semantic Web Technologies

Semantic Analysis of Multimedia Documents

* Content-Based Multimedia Analysis
* Knowledge Assisted Multimedia Analysis and Data Mining

Semantic Retrieval of Multimedia Documents

* Semantic-Driven Multimedia Indexing and Retrieval
* Machine Learning and Relevance Feedback for Semantic Retrieval
* Semantic-Driven Multimedia Content Adaption and Summarization

Linked Data and Multimedia

* Data Integration for Multimedia Documents
* Named Entity Recognition and Disambiguation of Linked Data Entities in 
   Multimedia Documents

Semantic User Interfaces for Multimedia Documents

* Human-Computer Interfaces for Multimedia Data Access
* Content Organization of Multimedia Documents
* Smart Visualization and Browsing of Multimedia Documents
* Visualization of Structured, Linked and Aggregated Data, Originating
   from Multiple Sources
* Ontology-based Interaction with Collections of Multimedia Data

Semantic Annotation and Tagging of Multimedia Documents

* User Generated Semantic Metadata for Multimedia Documents
* Interfaces and Personalization for Interaction and Annotation of
   Multimedia Documents

Semantic Metadata Management for Multimedia

* Metadata Management for Multimedia Documents
* Bridging Multimedia and Knowledge Domains
* Semantic-Driven Data Integration/Fusion of Media Streams in Multimedia
   Documents

Applications of Semantic Multimedia

* Semantic Multimedia Mash-ups
* Semantic-Driven Multimedia Applications in eLearning
* Semantic-Driven Multimedia Applications in Cultural Heritage Contexts


EDITORIAL REVIEW COMMITTEE
Tobias Buerger Capgemini, sdm, Germany
Fabio Ciravegna, U. of Sheffield, England
Thierry Declerck, DFKI Saarbr¸cken, Germany
Sigfried Handschuh, DERI, National U. of Ireland
Lynda Hardman, CWI Amsterdam, The Netherlands
Wolfgang Huerst, Utrecht U., The Netherlands
Ruediger Klein, Fraunhofer IAIS, Germany
Steffen Lohmann, U. Carlos III de Madrid, Spain
Matthias Lux, Alpen-Adria U. Klagenfurt, Austria
Antonio Penta, U. of Southampton, England
Tobias Thelen, U. of Osnabrueck, Germany
Raphaël Troncy, EURECOM, France


SUBMISSIONS

Authors are invited to submit high quality articles in IJSC formatting to:
https://www.easychair.org/conferences/?conf=smijsc2011

The guidelines for contributions including formatting templates for MS Word and 
Latex can be found at
http://www.worldscinet.com/ijsc/mkt/guidelines.shtml

In case of questions, please contact the guest editors at
robert.mert...@iais.fraunhofer.de
or
harald.s...@hpi.uni-potsdam.de


IMPORTANT DATES

* June 3, 2011: Submissions due
* DEADLINE EXTENSION: June 24, 2011: Submissions due
* August 3, 2011: Notification date
* October 18, 2011: Final versions due



Dr. Harald Sack
Hasso-Plattner-Institut für Softwaresystemtechnik GmbH
Prof.-Dr.-Helmert-Str. 2-3
D-14482 Potsdam
Germany
Amtsgericht Potsdam, HRB 12184
Geschäftsführung: Prof. Dr. Christoph Meinel
Tel.: +49 (0)331-5509-527
Fax: +49 (0)331-5509-325
E-Mail: harald.s...@hpi.uni-potsdam.de
http://www.hpi.uni-potsdam.de/meinel/team/sack.html