[Nottingham] Downloading multiple images from a website

Michael Quaintance penfoldq at penfoldq.co.uk
Tue May 31 23:47:14 BST 2005


Hi all

I'm trying to steal a load of images from a website I found and I'm 
struggling to work out how to do it fairly. (Alright, there's a 
contradiction there but, so what).

Basically, this site has a folder of images that are not directly 
linked from anywhere useful but are randomly chosen to fill in part of 
a php script for the 404 ErrorDocument on the site. Looking at the php 
output, I have found the directory in which they are held.

I want to steal all the files from (not the actual website)

http://www.website.com/images/random404img/

with the filenames 0.jpg to 999.jpg but acknowledging that not all of 
these will exist and not wanting to hammer his bandwidth so I want to 
make sure that I wait a reasonable amount of time between each one.

The website doesn't have a robots.txt so I'm not trying something he 
has specifically disallowed.

It seems to scream for a simple shell script and something like wget 
but I have not been able to find the right options. Any ideas?

Cheers

-Penfold




More information about the Nottingham mailing list