The other option that comes to mind is to use rsync if you can. If the file isn't complete, next time it runs, it will get the rest of it. You could combine this with a cron job using find again to delete ones that are old. As long as you don't use rsync with --delete it won't delete your local copy if a file is missing on the server.
If you don't have rsync, then ncftpget, as John Turner mentioned, is probably a good option.
Joel
Mark Freeze wrote:
A year or so ago I had a problem downloading a file via ftp onto a Windows box with WS_FTP. The file was about 100MB and I started downloading the file while my customer was still uploading, so I only got about half of the file. WS_FTP allowed me to do this with no error. (Which I thought was kinda crazy.)
Now I have an offsite ftp spot that my customers use to send me files at random times during the day. I want to automatically download and process these files onto my box as soon as they appear on the site so I was thinking that I would scehedule up a cron job to look for these files every 10 min. When I do this am I going to have the problem of seeing the file and trying to get it as they are uploading? Some of these files are over 100MB and might take my customer a while to upload. Someone told me to make sure that I have exclusive access to the file before I download it, but since I have no control over the ftp server I'm not sure on how to accomplish that task.
Any help will be greatly appreciated.
Regards, Mark.
-- TriLUG mailing list : http://www.trilug.org/mailman/listinfo/trilug TriLUG Organizational FAQ : http://trilug.org/faq/ TriLUG Member Services FAQ : http://members.trilug.org/services_faq/ TriLUG PGP Keyring : http://trilug.org/~chrish/trilug.asc
