Using wget to snapshot a web site

Wil McGilvery wmcgilvery
Wed Dec 1 12:06:12 PST 2004


I have used spiderzilla as a plug in for firefox. It is a front end for htdig I believe.

Regards,

Wil McGilvery
Manager
Lynch Digital Media Inc

         

905-363-1600
905-363-4297 Ext. 248
416-716-3964 (cell)
1-866-314-4678
905-363-1194? FAX
www.LynchDigital.com


-----Original Message-----
From: linux-users-bounces at linux-sxs.org [mailto:linux-users-bounces at linux-sxs.org] On Behalf Of Michael Hipp
Sent: Wednesday, December 07, 2005 4:01 PM
To: Linux tips and tricks
Subject: Using wget to snapshot a web site

I'm trying to use wget to grab an offline copy of this website so I can 
refer to it when doing development without Internet access.

    http://wiki.wxpython.org/index.cgi/FrontPage

But all the links in that page all look like this:

    <a href="/index.cgi/ObstacleCourse">ObstacleCourse</a>

I can't find any combination of options for wget which will cause it to 
follow these links. I presume it's because the link is written like an 
absolute link when it is actually more of a relative link.

Anyone know how to get wget to grab these or another tool which might do 
the job?

Thanks,
Michael
_______________________________________________
Linux-users mailing list
Linux-users at linux-sxs.org
http://mail.linux-sxs.org/cgi-bin/mailman/listinfo/linux-users

Need to chat further on this subject? Check out #linux-users on irc.linux-sxs.org !




More information about the Linux-users mailing list