Hi ya. I've just put together a script file that will call wget to backup one of the websites I manage, and built a cron job to run this every night. The problem I'm having is in accessing a specific ftp folder.
The host the site is with has assigned our user account a default directory, so when I log in my path is something like "ftp.mysite.com/html". There is a database file that is one level above this "ftp.mysite.com/database" (assuming I got the path right). The datbase folder has restricted access so that web users cannot download the database file, or even see the folder. When I use a tool like SmartFTP, I can access both folders without having to login. However, when I open up a command line ftp session on my linux box, I cannot connect to the database folder - I get permission denied, even though I'm logged in with the correct user account. Any thoughts on this? The command I'm feeding into wget right now to backup the html files is this: wget -N -r -i source_files.txt My source_files.txt contains these lines: ftp://username:[EMAIL PROTECTED]/ ftp://username:[EMAIL PROTECTED]/database The second line fails. The problem is connecting to the database folder so I can backup the database file. Once I'm able to do this, I can apply it to wget, and backup the site properly. I guess an alternative would be to use an FTP script instead of wget, but the connection problems still exists for the database folder. If it helps any, the host is running Serv-U FTP-Server v2.5m Thanks for any suggestions/tips. It feels like I've missed something small somewhere, but I'm not seeing it right now..... Shawn
