On Thu, Dec 3, 2009 at 2:43 AM, George De Bruin <sndcha...@gmail.com> wrote:
> Actually, I was thinking about the ScrapBook extension when I started
> reading this thread... The thing that the ScrapBook extension could be is a
> model of what a Zim plugin should look like... The biggest thing is the
> ability to (a) scan X layers deep into a page, and (b) replace links with
> the local cached versions of the images / pages / etc, (c) the ability to
> selectively ignore sets of links, and (d) the ability to replace links that
> don't have pages / images in the cache with dummy links.
> IMO - this kind of functionality is really a bit more complicated than it
> appears at first blush.
Well, it doesn't have to be too complicated as long as you don't try to do
everything yourself. Downloading webpages can be done by a specialized
program like wget, which will also adopt all the links in the downloaded
pages to local links for you and has many other options.
Searching through a zim notebook and getting urls to feed to wget is
trivial, just open a page, iterate link elements in the parse tree and list
all links that start with "http" (or a short list of supported protocols).
Only part I would have to look into as where to place the hooks that allow
links to go through a plugin instead of to the browser directly when opening
a link in zim. But this should not be hard - just an implementation choice.
Mailing list: https://launchpad.net/~zim-wiki
Post to : email@example.com
Unsubscribe : https://launchpad.net/~zim-wiki
More help : https://help.launchpad.net/ListHelp