[Dub-pm] big data structures relative to memory size
seanpor at acm.org
Fri Apr 16 09:08:23 CDT 2004
I've an analysis program with a couple of million records that i really
need to keep in memory as i need to scan back and forth etc... With 5
million odd records (written as a couple of independent 'arrays' or
should i say 'lists') the program requires quite a bit more than the
1.5Gb of ram and becomes very slow due to swapping - gentoo-linux...
Each record has 5 integers and a string of max.len 30 chars... but perl
takes up extra ram for each SV... I would like to be able to handle
larger datasets much faster than currently...
Has anybody used INLINE::C for handling large data structures - if so
how do you load the info?
Anybody used PDL?
Any thoughts which way I should jump?
More information about the Dublin-pm