On Tue, 16 Jun 2009 11:51:10 +0200 Patrik Lembke <[email protected]> wrote:
> On Mon, 15 Jun 2009 22:00:11 +0930 Karl Goetz <[email protected]> > wrote: > > I would prefer something that runs server side and reads PMWikis > > data files. Perhaps then emailing somewhere (this list once a > > week?) or dropping it in a tarball somewhere. > > Well that would be a lot of work (either reimplementing the parser of > pmwiki or somehow interfacing that part of pmwiki's code). Unfortunate but true. > > My idea was that of the spider was that it only should be used > (widely) by some gNewsense team in a maner similar to what you > proposed. I'm worried about needlessly hitting up the http server by crawling it. > > > We have been discussing changing the webapps we use to run the > > project, so the wiki format may change in the future. Unfortunately > > theres been no time to effectively test the new webapps to > > determine whats useful. kk > > In that case writing a parser for a specific wiki-engine is even more > of a unnecessary and hard task. Again, unfortunate but true. > > So in effect I choose to write a spider (that I have written most of > only last night). > > A sample of the current implementation could be seen here: > http://blambi.chebab.com/dump/wikidoc-sample/ > > Its currently only of depth one (it just downloads pages that > are linked on HomePage), and links to non-english pages wont work yet. Let us know how you progress. kk > PS. sorry about no signing, I've managed to break gpg. -- Karl Goetz, (Kamping_Kaiser / VK5FOSS) Debian contributor / gNewSense Maintainer http://www.kgoetz.id.au No, I won't join your social networking group _______________________________________________ gNewSense-dev mailing list [email protected] http://lists.nongnu.org/mailman/listinfo/gnewsense-dev
