On Wed, Jan 23, 2019 at 12:54:51PM -0500, Ben Gamari wrote: > > > Sounds like we are largely in agreement. Let's start on this after the > Trac migration is finished.
I can in fact start working on this already while the migration is pipelined. It's just markdown in git, so writing a draft and then merging it into the wiki is perfectly straightforward. > I was going to recommend we rather just use something off-the-shelf to > avoid reinventing the wheel. I use linkchecker [1] elsewhere and it > works quite well. I believe you can use its -W flag to catch > non-existent wiki pages (which won't return a 404). Ofc, if we can do it without any extra coding effort, all the better. I do think doing it as part of the migration script might make it easier to capture those "redirect" pages (e.g. https://ghc.haskell.org/trac/ghc/wiki/Newcomers): we're scraping anyway, so I think detecting the shape of those pages shouldn't be overly difficult, and might be something an off-the-shelf link checker isn't as suitable for. _______________________________________________ ghc-devs mailing list [email protected] http://mail.haskell.org/cgi-bin/mailman/listinfo/ghc-devs
