On 26 January 2014 19:38, David Gerard <[email protected]> wrote:

> On 27 January 2014 00:17, Kevin Gorman <[email protected]> wrote:
>
> > Funny you ask... there are not currently any solid ones afaik, but I've
> > been talking with the Internet Archive about building out a bot and
> trying
> > to achieve community consensus on ENWP to autoreplace deadlinks with
> > archive.org ones.  The IA has been crawling all new external links on
> all
> > Wikimedia projects at least once every couple of hours for months, and
> has
> > a strong interest in killing off literally all of our dead links.  Unless
> > something falls through, I should be bringing a more detailed plan up
> > within maybe five or six weeks.
>
>
> Yes, I knew you were cooking up something :-) I was just surprised it
> wasn't the sort of task that people had already automated, or written
> a nice toolserver bot for, or something.
>
> The ones that use {{cite web}} and variants are pretty simple: you
> just whack in archiveurl= and archivedate= (preferably as close as
> possible to any cited accessdate=) ... then double-check by eye, of
> course. It just gets very tedious and error-prone doing it by hand,
> cut'n'pasting URLs into the middle of the computer guacamole we
> lovingly euphemise as "wikitext". VE isn't a much happier method.
>
>
Concur that it's a great idea....but perhaps a WMF Tools labs tool, instead
of toolserver?  Running battle, I know - but so many of the tools I have
greatly valued over the years are now pretty much useless, or at least
unreliable.

In any case - it would be great to have a bot that did a fair bit of that,
but it should probably be manually run to ensure proper matching, kind of
like AWB.

Risker/Anne
_______________________________________________
WikiEN-l mailing list
[email protected]
To unsubscribe from this mailing list, visit:
https://lists.wikimedia.org/mailman/listinfo/wikien-l

Reply via email to