Using wget to snapshot a web site
Wil McGilvery
wmcgilvery
Wed Dec 1 12:06:12 PST 2004
I have used spiderzilla as a plug in for firefox. It is a front end for htdig I believe.
Regards,
Wil McGilvery
Manager
Lynch Digital Media Inc
905-363-1600
905-363-4297 Ext. 248
416-716-3964 (cell)
1-866-314-4678
905-363-1194? FAX
www.LynchDigital.com
-----Original Message-----
From: linux-users-bounces at linux-sxs.org [mailto:linux-users-bounces at linux-sxs.org] On Behalf Of Michael Hipp
Sent: Wednesday, December 07, 2005 4:01 PM
To: Linux tips and tricks
Subject: Using wget to snapshot a web site
I'm trying to use wget to grab an offline copy of this website so I can
refer to it when doing development without Internet access.
http://wiki.wxpython.org/index.cgi/FrontPage
But all the links in that page all look like this:
<a href="/index.cgi/ObstacleCourse">ObstacleCourse</a>
I can't find any combination of options for wget which will cause it to
follow these links. I presume it's because the link is written like an
absolute link when it is actually more of a relative link.
Anyone know how to get wget to grab these or another tool which might do
the job?
Thanks,
Michael
_______________________________________________
Linux-users mailing list
Linux-users at linux-sxs.org
http://mail.linux-sxs.org/cgi-bin/mailman/listinfo/linux-users
Need to chat further on this subject? Check out #linux-users on irc.linux-sxs.org !
More information about the Linux-users
mailing list