Vinh Nguyen wrote: > Dear list, > > My goal is to download some pdf files from a dynamic site (not sure on > the terminology). For example, I would execute: > > wget -U firefox -r -l1 -nd -e robots=off -A '*.pdf,*.pdf.*' > http://site.com/?sortorder=asc&p_o=0 > > and would get my 10 pdf files. On the page I can click a "Next" link > (to have more files), and I execute: > > wget -U firefox -r -l1 -nd -e robots=off -A '*.pdf,*.pdf.*' > http://site.com/?sortorder=asc&p_o=10 > > However, the downloaded files are identical to the previous. I tried > the cookies setting and referer setting: > > wget -U firefox --cookies=on --keep-session-cookies > --save-cookies=cookie.txt -r -l1 -nd -e robots=off -A '*.pdf,*.pdf.*' > http://site.com/?sortorder=asc&p_o=0 > wget -U firefox --referer='http://site.com/?sortorder=asc&p_o=0' > --cookies=on --load-cookies=cookie.txt --keep-session-cookies > --save-cookies=cookie.txt -r -l1 -nd -e robots=off -A '*.pdf,*.pdf.*' > http://site.com/?sortorder=asc&p_o=10 > > but the results again are identical. Any suggestions? > > Thanks. > Vinh
Look at the page source how they are generating the urls. Maybe they are using some ugly javascript, although that discards the benefit of paging...