FYI:
If anyone used HTTrack, a common problem is it will download the webpages, but on certain websites it wont download the images.
Websites are complicated thingies behind the scenes, so HTTrack has complicated options to fix that (and its got lousy help texts imo).
But a simple method Ive found is go to "Set Options" --> Spider, and disable the Robots.txt file.
Works for me, maybe it will work for you.
Wednesday, August 6, 2008
Subscribe to:
Post Comments (Atom)
4 comments:
Try out wget, if you want something a bit more mature. It is definitely the gold standard, and should have plenty of documentation, although it is entirely command line...
Anon,
Ill check it out. Thanks!
Let me go for apply first…
Pretty girl gets off wonder what would have happened if she had been older and a bit plain …
Post a Comment