[Nh-pm] How to save complete web page not just text?
James Kellndorfer
jameskel at adelphia.net
Fri Jun 24 14:10:00 PDT 2005
I got wget to work properly.
As far as I'm concerned, using wget suits my needs. Problem solved.
Thanks,
JK
----- Original Message -----
From: "Kevin D. Clark" <clark_k at pannaway.com>
To: "James Kellndorfer" <jameskel at adelphia.net>
Cc: <nh-pm at pm.org>
Sent: Friday, June 24, 2005 2:16 PM
Subject: Re: [Nh-pm] How to save complete web page not just text?
>
> "James Kellndorfer" <jameskel at adelphia.net> writes:
>
> > The webpage source that you retrieved using LWP will not contain the .js
> > files that contain the data for each 5 minute period.
>
> $ wget -E -H -k -K -p http://marketrac.nyse.com/ot/ordertrac_detail.html
>
> ....
>
> FINISHED --14:11:41--
> Downloaded: 69,000 bytes in 37 files
> Converting marketrac.nyse.com/ot/ordertrac_detail.html... 50-3
> Converting www.nyse.com/404.html... 1-0
> Converted 2 files in 0.00 seconds.
> $ ls
> marketrac.nyse.com/ www.nyse.com/
> $ find . -name \*.js
> ./marketrac.nyse.com/ot/_rnd.js
> ./marketrac.nyse.com/ot/_locate.js
> ./marketrac.nyse.com/data/siac/OrderTrac/staticFiles/js/snapshot.js
> ./marketrac.nyse.com/data/siac/OrderTrac/staticFiles/js/NYA.js
> ./www.nyse.com/redirect.js
> $
>
> Aren't those what you want? snapshot.js and NYA.js have a lot of data
> in them.
>
> Regards,
>
> --kevin
> --
> GnuPG ID: B280F24E And the madness of the crowd
> alumni.unh.edu!kdc Is an epileptic fit
> -- Tom Waits
More information about the Nh-pm
mailing list