Hey Joshua,

On Aug 15, 2006, at 1:33 PM, Joshua Schachter wrote:

> All kinds of subtlety here. For example, what to do if the site  
> happens
> to be down while we check it? What about respecting robots.txt etc?

You could explicitly only use HTTP return codes. That is to say, only  
take action for sites that are accessible, and return certain codes,  
eg: 301 (permanently moved), 404 (not found), and 410 (gone). A site  
outage would (hopefully) be returning 500 or nothing at all. Sites  
that weren't accessible (DNS, no route, etc) would be left alone.

I'm not sure what to do about robots.txt, which also raises a more  
fundamental question of whether d.i.u really should be actively doing  
anything at all, as opposed to passively accepting submissions.  
Perhaps people who want additional features like this should just  
work towards client-side applications that use the d.i.u API, and  
leave the site itself as a repository.

- siege

---
Christopher (siege) O'Brien
[EMAIL PROTECTED]




 
Yahoo! Groups Links

<*> To visit your group on the web, go to:
    http://groups.yahoo.com/group/ydn-delicious/

<*> To unsubscribe from this group, send an email to:
    [EMAIL PROTECTED]

<*> Your use of Yahoo! Groups is subject to:
    http://docs.yahoo.com/info/terms/
 



Reply via email to