Grabbing An Entire Website
Aaron
aaron at rock.idev.com
Wed Apr 19 22:53:29 EDT 2000
yup, wget -r www.foobar.com. Of course that gets what a browser would
"see" not the source code behind dymanic pages, unless of course it's cold
fusion ;) If you want to get source code, for dynamic pages or something
else that would depend on the situation.
Aaron
On Wed, 19 Apr 2000, Garrett Nievin wrote:
> I think that you can use wget for that. Have not done it myself.
>
>
> Cheers,
> Garrett
>
> On Wed, 19 Apr 2000, Janina Sajka wrote:
>
> > Hi:
> >
> > Anyone know how to auto-retrieve an entire www page hierarchy?
> >
> > I know software like ncftp can and wuftp can tar up an entire directory
> > tree, but the pages I need aren't available over ftp, only http. I'd hate
> > to have them by hand one at a time, though.
> >
> >
>
> --
> Garrett P. Nievin <gnievin at gmu.edu>
>
> Non est ad astra mollis e terris via. -- Seneca
>
More information about the Speakup
mailing list