[Nh-pm] How to save complete web page not just text?

James Kellndorfer jameskel at adelphia.net
Fri Jun 24 14:10:00 PDT 2005


I got wget to work properly.

As far as I'm concerned, using wget suits my needs. Problem solved.

Thanks,
JK


----- Original Message ----- 
From: "Kevin D. Clark" <clark_k at pannaway.com>
To: "James Kellndorfer" <jameskel at adelphia.net>
Cc: <nh-pm at pm.org>
Sent: Friday, June 24, 2005 2:16 PM
Subject: Re: [Nh-pm] How to save complete web page not just text?


>
> "James Kellndorfer" <jameskel at adelphia.net> writes:
>
> > The webpage source that you retrieved using LWP will not contain the .js
> > files that contain the data for each 5 minute period.
>
> $ wget -E -H -k -K -p http://marketrac.nyse.com/ot/ordertrac_detail.html
>
> ....
>
> FINISHED --14:11:41--
> Downloaded: 69,000 bytes in 37 files
> Converting marketrac.nyse.com/ot/ordertrac_detail.html... 50-3
> Converting www.nyse.com/404.html... 1-0
> Converted 2 files in 0.00 seconds.
> $ ls
> marketrac.nyse.com/  www.nyse.com/
> $ find . -name \*.js
> ./marketrac.nyse.com/ot/_rnd.js
> ./marketrac.nyse.com/ot/_locate.js
> ./marketrac.nyse.com/data/siac/OrderTrac/staticFiles/js/snapshot.js
> ./marketrac.nyse.com/data/siac/OrderTrac/staticFiles/js/NYA.js
> ./www.nyse.com/redirect.js
> $
>
> Aren't those what you want?  snapshot.js and NYA.js have a lot of data
> in them.
>
> Regards,
>
> --kevin
> -- 
> GnuPG ID: B280F24E                     And the madness of the crowd
> alumni.unh.edu!kdc                     Is an epileptic fit
>                                        -- Tom Waits



More information about the Nh-pm mailing list