Is there a program that can save every image from an HTML file?

I have a longass HTML file with about a billion images linked from it. Is there a program that can save each of those images indivdually, so as to prevent me from having to right click each one and choose “Save file As”?

Do you mean you have a gigantic web page with a gazillion images in it, or do you mean you have a web site of several web pages with a gazillion images within the site?

If just a single web page with lots of images, Firefox does it for you no sweat. Just save the web page to your hard drive like this —> [ File | Save Complete Page As … ]

If you have images spread over several web pages, you need a tool such as Teleport, a web spider, to do the job. There are several apps that will work, including open source. Teleport was the one that first came to mind at the moment.

I wonder what would happen if you tried a tool like Teleport and told it to download everything at say “amazon.com” or “wikipedia.org”. Would the WWW Police show up at your door and slap you around for plugging up all the tubes of the Internet?

The DownThemAll Firefox extension may work.

I’ve actually attempted this with a web site about 250 MB in size. It puts considerable strain on the web server and if there’s an alert sys admin at the server farm they usually kill the process.

In Opera, you can do File|Save as… and choose “HTML file with images” as the file type.

It is an interesting question of the information age but why do you think someone like Google has any more right to grab everything than you as a person? It is great that they are allowed to do what they do but the non-digital equivalent would get someone in massive trouble.

“Excuse me miss, can I hold all those papers on your desk? photocopies everything and hands it back. Sticks the results all over public places for all to use. Thanks!”

Anyway, there is no way an individual can hose up any fraction of the internet. The best that you could do is piss off your ISP but that will take way more than hundreds of images in this day when everyday people run file sharing software all day long.

Grabbing things off of web sites for your own personal use, no matter how big it is, seems to be allowed. It really has to be for modern technology to work. People suck down content to use on the go and software often needs things for internal caching.

Duckster’s suggestion works equally well in IE 6 & 7.

Go to the page uo want to grab images from, select "Save as … " from the File menu (IE6) or Page menu (IE7). Then change the “Save as type” dropdown to “Webpage, complete (*.htm, *.html)”. Click [Save]

The result will be a page saved with whatever name & a folder named “whatevername_ files”. All your images will be in the folder.

Takes 3 clicks.

Not all of amazon.com (and not with Teleport) but just the CD covers and song titles for making jukebox labels.

Kills? Or just alert when Robot.txt is exceeded? Can you give an example of the alert system you mentioned?

WinHTTrack doesn’t seem to run into a file size limit, at least not that I’ve noticed. FTR though it pulls everything off the webpage, images, html pages, zips etc so you may need to do some deleting later.