Re: How?

2023-10-24 Thread Fernando Cassia
On Tue, 24 Oct 2023, 17:16 ENG WKJC,  wrote:

> Guys,
>
> Been looking for a solution that can run on Win7 or higher for HTTPS
> downloading.  However, I don't understand the whole GNU thing, other than
> it's
> open source and generally free software.
> I've looked at the software links and don't know what to do here.


HI Marv

You can find pre-built versions of wget for Windows on this Web site

32bit
https://eternallybored.org/misc/wget/1.21.4/32/wget.exe

And 64 bit
https://eternallybored.org/misc/wget/1.21.4/64/wget.exe

You just download the exe required for your system.

How to tell if your Windows installation is 32bit or 64bit
https://support.microsoft.com/en-us/windows/32-bit-and-64-bit-windows-frequently-asked-questions-c6ca9541-8dce-4d48-0415-94a3faa2e13d

... then manually copy to a folder in your computer, and then place that
folder in your "path" (the path is a listing of folder locations that
Windows uses to look for exe files)

(I personally use c:\utils and place all command line utilities there).

---
>From the command prompt:
MD C:\Utils
Copy %userprofile%\downloads\wget.exe c:\utils
---

Here is a tutorial on how to the add that folder you ju St created to the
system path

https://www.architectryan.com/2018/03/17/add-to-the-path-on-windows-10/

You can then run wget from the command prompt from any directory or folder
just by calling it with 'wget'

Eg to get a listing of all options available
wget - -help

Others will be able to guide you after that.

If you want to see a video, sometimes a picture is better than a thousand
words

https://youtu.be/cvvcG1a7dOM?si=Q5hO2GzlFFG0oRhC

Best,

FC
Buenos Aires, Argentina


Re: [Bug-wget] wget 1.11.4 ignoring -np

2013-12-03 Thread Fernando Cassia
On Mon, Dec 2, 2013 at 8:34 PM, Micah Cowan mi...@cowan.name wrote:
 You're really specifying www.somehost.com/mirrors/mozilla/extensions/ and
 not www.somehost.com/mirrors/mozilla/extensions (no final slash)? Because
 even though they may be the same page in reality, and the server may even
 redirect the latter to the former, wget would see the final extensions as
 a file, and the preceding /mirrors/mozilla/ as the actual dir, and so
 would happily spider anything provided that it begins /mirrors/mozilla. The
 final slash makes a big difference!

Thanks Micah.

I think you've hit the nail in the head. :) Thanks so much for your
time and response, and the info wrt the current maintainer.

Best regards!
FC


-- 
During times of Universal Deceit, telling the truth becomes a revolutionary act
Durante épocas de Engaño Universal, decir la verdad se convierte en un
Acto Revolucionario
- George Orwell



Re: [Bug-wget] Big refactoring

2012-08-07 Thread Fernando Cassia
On Tue, Aug 7, 2012 at 12:12 PM, illusionoflife
illusion.of.lif...@gmail.com wrote:
 I really want to contribute to Wget, but I am afraid that such radical changes
 will not be accepted with

I´m just a lurker on this list, and have no relation whatsoever with
the wget project -other than being a long-time user, I just want to
say that I´d feel much safer if contributors came from a real
identifiable person with a face and name and email address, rather
than from an anonymous entity hidden behind ´illusionoflife´. But hey,
that´s me.

FC

-- 
During times of Universal Deceit, telling the truth becomes a revolutionary act
Durante épocas de Engaño Universal, decir la verdad se convierte en un
Acto Revolucionario
- George Orwell



Re: [Bug-wget] Big refactoring

2012-08-07 Thread Fernando Cassia
On Tue, Aug 7, 2012 at 7:53 PM, Micah Cowan mi...@cowan.name wrote:
 Wget is monolithic, portable to non-Unix, written entirely in
 C, and can be built to have few dependencies. My Niwt project has the
 specific aims to be as hackable and behavior-changeable as possible,

Then I say both serve specific purposes... ie ´classic wget´ is more
easily portable (win32, os2, you-name-it), while next-gen wget would
be unix-linux oriented.

Why not just bump the version and let´s keep ´classic´ wget and ´wget
2.0´ in parallel?.
If ´classic´ wget rots someone else will come along and continue
enhancing it to support new protocols etc (spdy anyone?) because he
needs a current tool for his non-unix based OS...

FC



[Bug-wget] Can someone please help with wget log-in ? apparently I'm missing some Javascript?

2012-06-06 Thread Fernando Cassia
There is this site that has some files available for download.
http://ginga.lavid.ufpb.br/projects/openginga/files

For testing use
user:bugmenot
password:please

I am trying to automatize the download of one of the files via wget.
I' m using keep-session cookies along with --load-cookies and
--save-cookies before each request, yet what I'm getting back is
always the log-in form.

I think if you look at the source (log-in form) there' s a session
token there, apparently handled via Javascript.

Just to make sure the web server is not discriminating against wget, I
also used --user-agent matching that of Firefox 13 for Linux.

Here' s my script, and you'll see quickly that it doesn't work:.

/
#!/bin/bash
# User-agent: for FF13 is Mozilla/5.0 (X11; Linux i686; rv:13.0)
Gecko/20100101 Firefox/13.0
#
# get the log-in screen
wget --unlink --user-agent=Mozilla/5.0 (X11; Linux i686; rv:13.0)
Gecko/20100101 Firefox/13.0 --load-cookies=cookies.txt
--save-cookies=cookies.txt --keep-session-cookies --http-user=bugmenot
-http-passwd=please
http://ginga.lavid.ufpb.br/login?back_url=http%3A%2F%2Fginga.lavid.ufpb.br%2Fprojects%2Fopenginga%2Ffiles
#
# post user and password
wget --unlink --user-agent=Mozilla/5.0 (X11; Linux i686; rv:13.0)
Gecko/20100101 Firefox/13.0 --load-cookies=cookies.txt
--save-cookies=cookies.txt --keep-session-cookies --http-user=bugmenot
-http-passwd=please --http-post
http://ginga.lavid.ufpb.br/login?back_url=http%3A%2F%2Fginga.lavid.ufpb.br%2Fprojects%2Fopenginga%2Ffiles;
--post-date=username=bugmenotpassword=please
--referer=http://ginga.lavid.ufpb.br/login?back_url=http%3A%2F%2Fginga.lavid.ufpb.br%2Fprojects%2Fopenginga%2Ffiles;
#
# now I should have the log-in credentials, get the files list (just in case)
#
wget --unlink --user-agent=Mozilla/5.0 (X11; Linux i686; rv:13.0)
Gecko/20100101 Firefox/13.0 --load-cookies=cookies.txt
--save-cookies=cookies.txt --keep-session-cookies --http-user=bugmenot
-http-passwd=please
http://ginga.lavid.ufpb.br/projects/openginga/files;
#
# now get the file I want
#
wget --unlink --user-agent=Mozilla/5.0 (X11; Linux i686; rv:13.0)
Gecko/20100101 Firefox/13.0 -c --load-cookies=cookies.txt
--save-cookies=cookies.txt --keep-session-cookies
http://ginga.lavid.ufpb.br/attachments/download/475/deps-openginga-source-v0.4.3.tar.gz;
--output-document=deps-openginga-source-v0.4.3.tar.gz



Thanks in advance for any assistance you can provide.

FC

-- 
During times of Universal Deceit, telling the truth becomes a revolutionary act
Durante épocas de Engaño Universal, decir la verdad se convierte en un
Acto Revolucionario
- George Orwell



Re: [Bug-wget] Can someone please help with wget log-in ? apparently I'm missing some Javascript?

2012-06-06 Thread Fernando Cassia
On Wed, Jun 6, 2012 at 3:30 PM, Micah Cowan mi...@cowan.name wrote:

 Many CGI-based forms require all the data from the submitted form to be
 present in the post-data before they will function.


Thanks! great pointers. Will post back about any progress I make 

FC


Re: [Bug-wget] Support for long-haul high-bandwidth links

2011-11-30 Thread Fernando Cassia
2011/11/29 Andrew Daviel ad...@triumf.ca:
 but I became recently aware of download accelerators designed primarily to
 thwart bandwidth allocation/throttling. Interestingly Wget is listed on the
 Wikipedia page as a download manager, implying it can already do this.

 http://en.wikipedia.org/wiki/Download_acceleration

´Axel´ command line app

http://www.theinquirer.net/inquirer/news/1037769/don-download-accelerator

FC

-- 
The purpose of computing is insight, not numbers.
Richard Hamming - http://en.wikipedia.org/wiki/Hamming_code



Re: [Bug-wget] Support for long-haul high-bandwidth links

2011-11-30 Thread Fernando Cassia
On Wed, Nov 9, 2011 at 23:24, Andrew Daviel ad...@triumf.ca wrote:
 When downloading a large file over a high-latency (e.g. long physical
 distance) high-bandwidth link, the download time is dominated by the
 round-trip time for TCP handshakes.

Which is why large files should be stored on FTP servers, not http.

FTP was designed for a reason: to transfer large binary files reliably.

http was designed primarily to serve web pages, not large file downloads.

But try telling that to today´s sysadmins and webmasters educated aften Win95...

FC

-- 
The purpose of computing is insight, not numbers.
Richard Hamming - http://en.wikipedia.org/wiki/Hamming_code



Re: [Bug-wget] download that require login/password form submit

2010-04-09 Thread Fernando Cassia
On Fri, Apr 9, 2010 at 12:14 PM, Keisial keis...@gmail.com wrote:
 Voytek Eymont wrote:
 Micah,

 thanks !!
 I'm loging in OK.

 on next step I do like:

 wget --load-cookies=my-cookies.txt --save-cookies=my-cookies.txt
 --keep-session-cookies
 http://www.domain.tld/main.htm?_template=advanced_module=active_list

 that fails until I put  around the http string like so:

 wget --load-cookies=my-cookies.txt --save-cookies=my-cookies.txt
 --keep-session-cookies
 http://www.domain.tld/main.htm?_template=advanced_module=active_list;

 or should I use some '%' characters ? for  ? or just   around https
 string ?


 Just surround it with double   or single ' ' quotes.
 If  is not quoted your shell thinks you want to execute a program called
 wget and then assign active_list to a shell variable called _module (if
 there
 wasn't a = it would try to run a program called _module, which would give
 you an error message you could notice)


 next question: the resulting file has lots and lots of bumpf like
 space.gif galore, etc,

 how do I make into text as much as possible, is there a wget function, or ?

 Remove anything between  and , then unescape the entities. That should
 give you quite clean text with a minimal effort.

Use grep, and sed. Grep and sed are your friends

http://www.tech-recipes.com/rx/330/remove-html-tags-from-a-file/

FC