[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Re: webcrawl to cache dynamic pages



On Mon, May 09, 2005 at 10:31:08AM +0100, David Hugh-Jones wrote:
> 
> On 09/05/05, Richard Lyons <richard@the-place.net> wrote:
> > On Sun, May 08, 2005 at 09:48:07AM +0200, Nacho wrote:
> > > > On Mon, May 02, 2005 at 01:27:41PM +0100, Richard Lyons wrote:
> > > > > I am considering how to crawl a site which is dynamically generated,
> > > > > and create a static version of all generated pages (or selected
[...]
> > >
> > > Well, I don't know an "elegant" solution... one dirty approach would be to
> > > first download the site with "wget -r", then you would get lots of files with
[...]
> > 
> > Yes... that is the kind of thing I was imagining.  It will probably be
> > quite simple once I get started.  But first I need to find time :-(
[...]
> If you end up wanting to do something more complicated, you could look
> into WWW::Mechanize:
> 
> http://search.cpan.org/perldoc?WWW%3A%3AMechanize
>
That looks _very_ interesting --and potent.  Thanks for that too.

-- 
richard



Reply to: