[Nottingham] Downloading multiple images from a website
Michael Quaintance
penfoldq at penfoldq.co.uk
Tue May 31 23:47:14 BST 2005
Hi all
I'm trying to steal a load of images from a website I found and I'm
struggling to work out how to do it fairly. (Alright, there's a
contradiction there but, so what).
Basically, this site has a folder of images that are not directly
linked from anywhere useful but are randomly chosen to fill in part of
a php script for the 404 ErrorDocument on the site. Looking at the php
output, I have found the directory in which they are held.
I want to steal all the files from (not the actual website)
http://www.website.com/images/random404img/
with the filenames 0.jpg to 999.jpg but acknowledging that not all of
these will exist and not wanting to hammer his bandwidth so I want to
make sure that I wait a reasonable amount of time between each one.
The website doesn't have a robots.txt so I'm not trying something he
has specifically disallowed.
It seems to scream for a simple shell script and something like wget
but I have not been able to find the right options. Any ideas?
Cheers
-Penfold
More information about the Nottingham
mailing list