[Bug-wget] Option for Wget to download web pages as complete MHTML files.

2011-09-04 Thread mattw

Giuseppe,

I think an option for Wget to download web pages as complete MHTML files 
would be very useful.


I could see such a feature being used for an online service that allows 
people to publicly archive any web page. An addition to being of general 
use, this could also provide the online community a tool to combat 
propaganda- In cases where information is being scrubbed from media or 
informational web sites, people could easily save and share events as 
they were originally reported.


Regards,


Matthew Shockey
www.CreativeGraphicSolutions.BIZ
"Design, Web, and Print Services"
Phone 865-963-8409


Publish your photos in seconds for FREE
TRY IM TOOLPACK at http://www.imtoolpack.com/default.aspx?rc=if4





[Bug-wget] Wget should not ignore quota specifications for single files.

2011-09-04 Thread mattw

Giuseppe Scrivano,

I don't think Wget should ignore quota specifications for single files 
as it does. (See the "-Q" or "--quota" switch in the User Manual.)


If someone specifically indicates a quota from the command line, then 
this is obviously what they want to do (for whatever reason). On the 
other hand, if Wget is called from inside another program with a quota 
for a single file, it ought to be respected for predictability and 
program safety.


In my case, I am using Wget inside a Perl program for an online service, 
and would like to set the maximum download size on single files as a 
safety/performance precaution.


I've tried several work-arounds for this, all with no success. Wget 
simply refuses to follow quota specifications for single files no matter 
how Wget is invoked.


Respecting quotas for single files would be useful in other situations 
where Wget is called automatically from within a script.


(Wget is an awesome program in any case.)

Regards,


Matthew Shockey
www.CreativeGraphicSolutions.BIZ
"Design, Web, and Print Services"
Phone 865-963-8409


Send any screenshot to your friends in seconds...
Works in all emails, instant messengers, blogs, forums and social networks.
TRY IM TOOLPACK at http://www.imtoolpack.com/default.aspx?rc=if2 for FREE





[Bug-wget] Suggestion: An option for Wget to reset all command-line defaults.

2011-09-04 Thread mattw

Giuseppe,

I wonder if Wget needs an option like --resetdefaults=yes to reset any 
changes that may have been made in the ".wgetrc" file.


This way, if Wget is used within a batch file or script that is to be 
distributed for use on various platforms/systems, the developer of the 
script can be sure how Wget will be called.


Regards,


Matthew Shockey
www.CreativeGraphicSolutions.BIZ
"Design, Web, and Print Services"
Phone 865-963-8409


Share photos & screenshots in seconds...
TRY FREE IM TOOLPACK at http://www.imtoolpack.com/default.aspx?rc=if1
Works in all emails, instant messengers, blogs, forums and social networks.





[Bug-wget] Wget is not downloading background images.

2011-09-04 Thread mattw

Giuseppe,

Wget does not seem to be including images in web pages that are called 
with the CSS background specification. For example:


style=" background:url('images/cgsb-bg-menu-2.jpg')"

My test page is properly validated as HTML 4.01 Transitional, but 
neither of the following Wget commands will include the background 
images with the download:


wget -p http://www.creativegraphicsolutions.biz/index.html

wget -E -H -k -K -p http://www.creativegraphicsolutions.biz/index.html

wget -mrnp http://www.creativegraphicsolutions.biz

In these specific tests, I am using GNU Wget 1.11.4 on a Windows platform.

Regards,


Matthew Shockey
www.CreativeGraphicSolutions.BIZ
"Design, Web, and Print Services"
Phone 865-963-8409


Send any screenshot to your friends in seconds...
Works in all emails, instant messengers, blogs, forums and social networks.
TRY IM TOOLPACK at http://www.imtoolpack.com/default.aspx?rc=if2 for FREE





[Bug-wget] Wget should not ignore quota specifications for single files.

2011-09-04 Thread matthew

Giuseppe Scrivano,

I don't think Wget should ignore quota specifications for single files 
as it does. (See the "-Q" or "--quota" switch in the User Manual.)


If someone specifically indicates a quota from the command line, then 
this is obviously what they want to do (for whatever reason). On the 
other hand, if Wget is called from inside another program with a quota 
for a single file, it ought to be respected for predictability and 
program safety.


In my case, I am using Wget inside a Perl program for an online service, 
and would like to set the maximum download size on single files as a 
safety/performance precaution.


I've tried several work-arounds for this, all with no success. Wget 
simply refuses to follow quota specifications for single files no matter 
how Wget is invoked.


Respecting quotas for single files would be useful in other situations 
where Wget is called automatically from within a script.


(Wget is an awesome program in any case.)

Regards,


Matthew Shockey
www.CreativeGraphicSolutions.BIZ
"Design, Web, and Print Services"
Phone 865-963-8409





Re: [Bug-wget] Storing the URL in extended attributes

2011-09-04 Thread Petr Pisar
On Mon, Oct 18, 2010 at 12:48:14PM +0200, Petr Pisar wrote:
> On Mon, Oct 18, 2010 at 12:27:24PM +0200, Michelle Konzack wrote:
> > Am 2010-10-17 22:38:51, hacktest Du folgendes herunter:
> > > I created a patch to store the URL inside the user xattrs of the
> > > downloaded file; this way, its origin can be identified afterwards.
> > > 
> > > I uploaded the change to my Github account and attached the diff, and
> > > I am still working on portability issues, but I'd like to hear some
> > > opinions on this:
> > > 
> > > http://github.com/wertarbyte/wget/tree/xattrurl
> > 
> > I am right that this works only on Windos and not GNU/Linux?
> > 
> See my patch for wget
> . It works on
> GNU/Linux, IRIX and Darwin. Other option is to use libattr as the API is not
> standardized (it support IRIX and GNU/Linux only). FreeBSD has yet another
> API.
> 

Just if somebody is still interested (maybe new wget maintainer), rebased
patch against 1.13.3 is still available on .

-- Petr



pgpNWXhdvliTt.pgp
Description: PGP signature


Re: [Bug-wget] wget 1.13: FIONBIO does not exist on solaris

2011-09-04 Thread Giuseppe Scrivano
"Christian Jullien"  writes:

> When compiling gnutls.c on solaris 10 sparc with gcc 4.6.1
> I get an error on:
>   ret = ioctl (fd, FIONBIO, &one);
> because FIONBIO is undefined.
>  
> Adding: 
>  
> #include 
>
> Let:
> #ifdef F_GETFL
>   ret = fcntl (fd, F_SETFL, flags | O_NONBLOCK);
>
> to be used instead. It then compiles and correctly works.
>  
> Thank you to see how to include sys/fcntl.h conditionnally. I checked but it
> is not clear to me when and why you decide to include this system file.
>  
> I'll be glad to test new versions for you.

Thanks to have reported it.  We can assume sys/fcntl.h is always present
as gnulib will provide a replacement on systems where this file is
missing.

The change I am going to commit is simply:

=== modified file 'src/gnutls.c'
--- src/gnutls.c2011-08-30 14:43:25 +
+++ src/gnutls.c2011-09-04 10:43:35 +
@@ -48,6 +48,8 @@
 #include "ptimer.h"
 #include "ssl.h"
 
+#include 
+
 #ifdef WIN32
 # include "w32sock.h"
 #endif