Alvarez, Angelo CIV NAVPACMETOCCEN JTWC wrote:
> and execute commands for each file which was
I am looking for a script (perl, etc.) or application which can retrieve files
using https, monitor a server for new files,
We currently receive at least 1.3GB of data each day, so it should be robust
enough to handle that load. Thanks.
Wget supports the grabbing of files over https.
As for monitoring a server for new files, using
the -N option will allow you see if there are any
new files and download them.
To execute commands on the downloaded files, just
put wget into a script, download the file to a
"staging" area, then do what you need to the
And it works on NMCI or legacy. No installing
required... so you or your clients won't get into
trouble by running in on NMCI seats.
If you need an example, I can show you the script
I use to populate the SIPR webpage here... it is
rudimentary, but will give you a basic idea...
Webmaster / Unix Admin
NAVPACMETOC Center San Diego