To do this you can use a
System.Net.WebClient[
^]
- use it to download the page as a big string
- save the string in a file to the harddisk
- parse the string, using regular expressions, for images and what else you want.
- download the images, and what else you want
(If you want to get all the pages of the website you will need to also parse for hyperlinks and recursively download all of those too)
Please do keep in mind that the home page has a link to the home page!
To get around circles like that use a Dictionary to keep track of what you have and have not downloaded. A dictionary can contain a dictionary, that way you can save index.html in the first one, and add a dictionary for asp.net, and in the 2d dictionary save index.html again.
In the end you can then recursively print the dictionary to give you the sitemap.
Do keep in mind that this will generate a lot of traffic, and might not always be allowed by the website owners.
Hope this helps you on your way :-)