Hi Guys:
Here's the scenario:
A program written in perl reads a large file and creates a very large
hash (actually a hash of records, hash of hash) and then used that hash in a
lookup for other data processing. The hash has about 3,096,000 entries and is
incrementing day by day. Right now the perl process consumes 1 G of RAM using a
32 bit perl, and there's no problem (i.e. out-of-memory error).
Is there a way to put in into a disk hash that is using written in pure
perl (No DBMS or tools such as Berkeley DB)? I saw on Google people recommend
DBM::Deep. Anyone here uses DBM::Deep? I like to know if people uses DBM::Deep
for large disk hash and learn the performance of using it.
Thanks in advance.
Regards,
Ludwig
_________________________________________________
Philippine Linux Users' Group (PLUG) Mailing List
http://lists.linux.org.ph/mailman/listinfo/plug
Searchable Archives: http://archives.free.net.ph