Training, Open Source computer languages
PerlPHPPythonMySQLApache / TomcatTclRubyJavaC and C++LinuxCSS 
Search for:
Home Accessibility Courses Diary The Mouth Forum Resources Site Map About Us Contact
wget - grabbing web pages from command line

Posted by admin (Graham Ellis), 11 September 2003
A quick note on the wget utility ... great way to grab a copy of a web page from the command line and save it to your local disc. Also has recursive and mirror options.  Great for mirroring .html .gif and .jpg files - beware of scripts and .php and .jsp files, as you'll get the parsed version rather than the originals ...
           wget  -m  -w2    http://www.sheepbingo.co.uk/index.html

That example will mirror the named site, starting at index.html.  The -w2 is important because it puts a 2 second delay between each page rather than burning up all the bandwidth on what might be a large transfer!

Wget is included with Linux distributions ... otherwise find it at
ftp://ftp.mirror.ac.uk/sites/ftp.gnu.org/gnu/wget/ or alternatives listed at http://www.gnu.org/software/wget/wget.html if you would like a windows binary!



This page is a thread posted to the opentalk forum at www.opentalk.org.uk and archived here for reference. To jump to the archive index please follow this link.

You can Add a comment or ranking to this page

© WELL HOUSE CONSULTANTS LTD., 2014: Well House Manor • 48 Spa Road • Melksham, Wiltshire • United Kingdom • SN12 7NY
PH: 01144 1225 708225 • FAX: 01144 1225 899360 • EMAIL: info@wellho.net • WEB: http://www.wellho.net • SKYPE: wellho