On Wed, 26 Sep 2012, [email protected] wrote:

> On Wed, Sep 26, 2012 at 06:16:31PM -0400, Champoux wrote:
> > On Wed, Sep 26, 2012 at 03:43:26PM -0400, Robert P. J. Day wrote:
> > >   i have a text file containing over 100 URLs (one per line), and i
> > > want to produce that many PDF files of those URL pages, with output
> > > filenames simply 1.pdf, 2.pdf and so on -- that numbering is
> > > required.
> >
> >     $ perl -MLWP::Simple=mirror -ne'next if /^SKIP/; \
> >         mirror( $_ => $..".pdf" ); print "saving $..pdf\n"' < list
> >
> >     ?
>
>       Wait, I might have misunderstood. Are the original links
> not pdfs, but you want to convert them on the fly?

  exactly.  all the URLs are simply htm/html/shtml, and i want to:

  1) grab them
  2) convert to PDF
  3) save to sequentially numbered files

i'm effectively trying to do what you can do in the chromium browser
with "Print > File", saving as pdf.  i just don't want to do it
manually over 100 times.

>       If so, yeah, it's going to be a mite more complicated. I did
> something similar in the past, if it can help:
> http://babyl.dyndns.org/techblog/entry/contact-sheet

  ok, i'll check it out, thanks.

rday

-- 

========================================================================
Robert P. J. Day                                 Ottawa, Ontario, CANADA
                        http://crashcourse.ca

Twitter:                                       http://twitter.com/rpjday
LinkedIn:                               http://ca.linkedin.com/in/rpjday
========================================================================
_______________________________________________
Linux mailing list
[email protected]
http://oclug.on.ca/mailman/listinfo/linux

Reply via email to