The --spider option doesn't work?
Unless I'm mistaken the --spider option should cause wget to not actually download any files. However that doesn't seem to work (it still downloads them and stores them in the local filesystem). -- Peter Eriksson [EMAIL PROTECTED]Phone:+46 13 28 2786 Computer Systems Manager/BOFHCell/GSM: +46 705 18 2786 Physics Department, Linköping University Room: Building F, F203 SE-581 83 Linköping, Sweden http://www.ifm.liu.se/~peter
ScanMail Message: To Recipient virus found or matched file blocking setting.
ScanMail for Microsoft Exchange has taken action on the message, please refer to the contents of this message for further details. Sender = [EMAIL PROTECTED] Recipient(s) = [EMAIL PROTECTED]; Subject = New Roman Scanning Time = 04/24/2002 17:10:32 Engine/Pattern = 6.150-1001/269 Action on message: The attachment Yni.exe matched file blocking settings. ScanMail has taken the Deleted action. In einer für Sie bestimmten Nachricht wurde ein als gefährlich eingestufter Anhang geblockt oder es wurde ein Virus gefunden. Der Absender der Nachricht wird ebenfalls automatisch informiert. Als gefährlich eingestuft gelten u.A. alle ausführbaren Dateien wie z.B. *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. Wenn sie eine Datei mit entsprechender Endung verschicken oder empfangen wollen, komprimieren sie diese bitte zu einer *.zip-Datei mit Winzip. An attachment has been blocked which is classified as dangerous or a Virus has been found in the mail received by you. The sender of this mail was automatically informed. Among the attachments classified as dangerous are all executable files like *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. If you need to send or receive such an attachment you should compress it first into a *.zip archive by using Winzip.
ScanMail Message: To Recipient virus found or matched file blocking setting.
ScanMail for Microsoft Exchange has taken action on the message, please refer to the contents of this message for further details. Sender = [EMAIL PROTECTED] Recipient(s) = [EMAIL PROTECTED]; Subject = Worm Klez.E immunity Scanning Time = 04/24/2002 17:13:30 Engine/Pattern = 6.150-1001/269 Action on message: The attachment showthreaded[19].pif matched file blocking settings. ScanMail has taken the Deleted action. In einer für Sie bestimmten Nachricht wurde ein als gefährlich eingestufter Anhang geblockt oder es wurde ein Virus gefunden. Der Absender der Nachricht wird ebenfalls automatisch informiert. Als gefährlich eingestuft gelten u.A. alle ausführbaren Dateien wie z.B. *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. Wenn sie eine Datei mit entsprechender Endung verschicken oder empfangen wollen, komprimieren sie diese bitte zu einer *.zip-Datei mit Winzip. An attachment has been blocked which is classified as dangerous or a Virus has been found in the mail received by you. The sender of this mail was automatically informed. Among the attachments classified as dangerous are all executable files like *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. If you need to send or receive such an attachment you should compress it first into a *.zip archive by using Winzip.
Re: Feature request
Hi Frederic! I'd like to know if there is a simple way to 'mirror' only the images from a galley (ie. without thumbnails). [...] I won't address the options you suggested, because I think they should be evaluated by a developper/coder. However, as I often download galleries (and have some myself), I might be able to give you a few hints: Restricting files to be downloaded by a) file-name b) the directory they are in To a): -R*.gif,*tn*,*thumb*,*_jpg*,*small* you get the picture I guess (pun not intended, but funny nevertheless). Works quite well. To b): --reject-dir *thumb* (I am not sure about the correct spelling/syntax, I currently have neither wget nor winzip -or similar- on this machine, sorry!) It also seems these options are incompatible: --continue with --recursive This could be useful, imho. IIRC, you are correct, but this is intentional. (right?) You probably think of the case where during a recursive download, the connection breaks and a large file is only partially downloaded. I could imagine that this might be useful. However, I see a problem when using timestamps, which normally require that a file be downloaded, if sizes local/on the server do not match, or the date on the server is newer. How should wget decide if it needs to re-get or continue the file? You could probably to smart guessing, but the chance of false decisions persists. As a matter of fact, the problem is also existing when using --continue on a single file, but then it is the user's decision and the story is therefore quite different (I think). CU Jens -- GMX - Die Kommunikationsplattform im Internet. http://www.gmx.net
Re: Feature request
On Wed, Apr 24, 2002 at 07:21:06AM +0200, Frederic Lochon (crazyfred) wrote: Hello, I'd like to know if there is a simple way to 'mirror' only the images from a galley (ie. without thumbnails). Depends on what you'd call easy I guess, apart from Jens' suggestion, you might want to take a look at piping the output to stdout and do various voodoo with it and then spawn another wget to read the output, something like... (using GNU's grep, really scary people could just use perl -e 'while(){something($_);}' or sed/awk or whatever fetish they have) $ wget -qO- http://www.somedomain.com/gallery.html | grep -i href= | \ grep -Ei jpg|jpeg|png | wget -Fi- -B http://www.somedomain.com/ ...might just work if all you want is the mirror only images that are a link, without having to download the thumbnails (this is just a silly example and far from perfect in any way, but you get the idea). However this might only be feasible in a *nix environment, I have no idea how That Other Popular OS would work in this matter but I doubt they bundle it with GNU grep. ... PS: sorry if these questions are recurent, I don't have time to read every post (on the news server at sunsite.dk) and I haven't seen any FAQ. I don't think there is one, but perhaps a FAQ-page with the most common (and not so common) questions that popup on the list from time to time might be a good idea? Thanks, Frederic Lochon -- Henrik van Ginhoven [EMAIL PROTECTED] Storage for all kinds of waste http://mise.wox.org ...including humans 9799-5
Re: add tar option
Herold Heiko [EMAIL PROTECTED] writes: But if I understand this correctly (sorry, sources not checked, foot in mouth ecc.) with -k wget still needs to correct the html files later, when it knows what has been downloaded and what not. So it can't print the file as soon as downloaded, only at the end. You are correct. `-k' would be at odds with any kind of streaming simply because it needs to process all the HTMLs after the fact. Regardless of other things, that breaks the streaming. On technical grounds, it might be hard to shoehorn Wget's mode of operation into what `tar' expects. For example, Wget might need to revisit directories in random order. I'm not sure if a tar stream is allowed to do that. A simple |sort should fix that, You're misunderstanding me. In my thought experiment, I meant that Wget output might be a tar stream itself, not a list of file names to feed to `tar'. I agree with the idea do one thing and do it well, after all we are not talking about a windows gui try-to-do-everything program here. Either I did not understand you correctly, or a simple list of files should be enough for every case. You didn't understand me, but you proposed something far better. Yes, something equivalent to `find''s `-print'/`-print0' would actually help the original poster. Or did you mean something else with serialization of the result ? A tar stream is an example of serialization of a set of files and directories. It turns an on-disk structure into a stream of bytes that can be transferred over a pipe or a network in order to re-create something resembling the original structure.
Re[2]: Feature request
It also seems these options are incompatible: --continue with --recursive This could be useful, imho. JR How should wget decide if it needs to re-get or continue the file? JR You could probably to smart guessing, but the chance of false decisions JR persists. Not wanting to repeat my post from a few days ago (but doing so nevertheless) the one way without checking all files online is to have wget write the downloaded file into a temp file (like *.wg! or something) and renaming it only after completing the download. Then it could be run with -nc and continue with the temp file which isn't renamed yet when the download is interrupted. But somebody smarter than me must see if this can be implemented and consider the implications. I have no idea. -- Brix
ScanMail Message: To Recipient virus found or matched file blocking setting.
ScanMail for Microsoft Exchange has taken action on the message, please refer to the contents of this message for further details. Sender = [EMAIL PROTECTED] Recipient(s) = [EMAIL PROTECTED]; Subject = Re:wget,how are you Scanning Time = 04/24/2002 22:21:09 Engine/Pattern = 6.150-1001/269 Action on message: The attachment out.exe matched file blocking settings. ScanMail has taken the Deleted action. In einer für Sie bestimmten Nachricht wurde ein als gefährlich eingestufter Anhang geblockt oder es wurde ein Virus gefunden. Der Absender der Nachricht wird ebenfalls automatisch informiert. Als gefährlich eingestuft gelten u.A. alle ausführbaren Dateien wie z.B. *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. Wenn sie eine Datei mit entsprechender Endung verschicken oder empfangen wollen, komprimieren sie diese bitte zu einer *.zip-Datei mit Winzip. An attachment has been blocked which is classified as dangerous or a Virus has been found in the mail received by you. The sender of this mail was automatically informed. Among the attachments classified as dangerous are all executable files like *.exe, *.bat, *.com, *.cmd, *.pif, *.scr. If you need to send or receive such an attachment you should compress it first into a *.zip archive by using Winzip.
Re: Feature request
Hi Brix! It also seems these options are incompatible: --continue with --recursive [...] JR How should wget decide if it needs to re-get or continue the file? [...] Brix: Not wanting to repeat my post from a few days ago (but doing so nevertheless) the one way without checking all files online is to have wget write the downloaded file into a temp file (like *.wg! or something) and renaming it only after completing the download. Sorry for not paying attention. It sounds like a good idea :) But I am no coder... CU Jens
Aviso de Virus
El Mensaje [EMAIL PROTECTED] enviado por poechita [EMAIL PROTECTED] (con dirección de respuesta: [EMAIL PROTECTED]) y enviado a [EMAIL PROTECTED] NO PUEDE LLEGAR A SU DESTINO debido a que se detecto la presecia de un virus en su interior. Los detalles del virus detectado se presentan a continuación: Actulize su antivirus y revise su equipo. - eTrust EZ Antivirus Version 5.4.2.0 Started scanning: 11:58:40 PM, 24/04/02 Major dat file v1000 Minor dat file v1987 Macro dat file Apr 18 2002 (VMD Ver 1.6) Scanning file(s)... AVPA246.tmp - Win32.Klez.H worm. Finished scanning: 11:58:40 PM, 24/04/02 Number of files scanned: 1. Number of infections: 1 Number of infected files not cleaned/deleted/renamed: 1 AVPA246.tmp (Win32.Klez.H worm) - Original Headers Text: - Received: from Mbeuqlft (unverified [207.249.181.149]) by tepa.com.mx (EMWAC SMTPRS 0.83) with SMTP id [EMAIL PROTECTED]; Wed, 24 Apr 2002 23:58:30 -0600 Date: Wed, 24 Apr 2002 23:58:30 -0600 Message-ID: [EMAIL PROTECTED] From: poechita [EMAIL PROTECTED] To: [EMAIL PROTECTED] Subject: Re:let's be friends MIME-Version: 1.0 - Anti-Viral Plugin for EMWAC Internet Mail Server (IMS): http://www.zaxalon.com/avp.html?ref=AVP201message?refby=tepa.com.mx