Showing posts from August, 2013

Website Offline Copy

Ever wanted to have an entire website available as a local copy? I did yesterday. There's a website with a forum that is about to be closed and that contains lots of useful info. So I thought: "Why not save it?" I gave some tools a try but I could not get them to get all the files. Then I decided to use the good old wget. I logged in on an old linux machine and after reading the man page for a while I used the following command to download the site: $ wget \ --recursive \ --no-clobber \ --page-requisites \ --html-extension \ --convert-links \ --restrict-file-names=windows \ --domains \ --no-parent \ www. In case the part of the site is protected with a log on page you may use cookies to access the content. Firstly get the cookie with the following command:  $ wget \           --save-cookies cookies.txt \           --post-data 'user=yourusername&password=yourpassword&#