The best thing to do is just try each one and see which one works best for what you want to do. You won't break anything... unless you Google for Google on google.com :p then you break the internets and people will be sad
As for php sites it will most likely save the files as html and change the reference to that file from .php to .html. You certainly aren't going to get the actual php code though, just the output of that code. On Fri, Jan 16, 2009 at 11:18 AM, Paul Saenz <[email protected]>wrote: > http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/ > > > This looks like something I've been looking for. > I could modify the HTML on a webpage so I can > navigate the website from a file on my local machine, > (even when it's not connected to the internet of course), > but that seems like a lot of work, especially if you change > every link. > > I was looking at the comments (below) and it looks like wget > can do the exact same thing with the -k option, but it isn't entirely > clear. > > Can anyone tell me if wget can do everything that this HTTrack > is supposed to do. > > Also, obviously it can't download any php script on a website, > so it almost seems useless anyways for sites that use php. > > Finally, when I try to navigate the the HTTrack site link, I get > a connection interrupted message. I'm wondering if LSU is > blocking the site. Can anyone get this site from their connection? > > Any comments? > Thanks > Paul > > > 1. Saeid Zebardast <http://zebardast.wordpress.com/> Says: > December 16th, 2007 at 5:50 > am<http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/#comment-32473> > Thanks. > you can use wget command for download a website: > wget URL -k -c -r -p > 2. fwiffo Says: > December 16th, 2007 at 9:55 > am<http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/#comment-32497> > Thanks, > this one is really useful. > 3. ton Says: > December 16th, 2007 at 11:07 > am<http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/#comment-32504> > "In > contrast, the recursive mirror function of Wget will not rearrange the > hyper-links on the web pages you downloaded, so they might still be > pointing > to remote locations." > this is what -k switch of wget is for. > 4. Evan "JabberWokky" Edwards <http://cheshirehall.org/> Says: > December 16th, 2007 at 12:28 > pm<http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/#comment-32514> > I'm > assuming that you didn't explain the difference between wget -k and this. > Or > did you just not know about the -k switch? > A bit more information would be nice… there is room for improvement > with wget, as it misses css includes and such, so this might be a better > option, but the only thing you cite as a difference actually isn't > different. > 5. visit Says: > August 25th, 2008 at 8:46 > am<http://debaday.debian.net/2007/12/16/httrack-website-crawler-copier/#comment-91356> > good > work, will back soon, great site congratulation!! > > > > > > ------------------------------ > Windows Live™: Keep your life in sync. Check it > out.<http://windowslive.com/explore?ocid=TXT_TAGLM_WL_t1_allup_explore_012009> > > _______________________________________________ > LinuxUsers mailing list > [email protected] > http://socallinux.org/cgi-bin/mailman/listinfo/linuxusers > > -- Peter Manis (678) 269-7979
