Hi all. I have the need to store a large (10M) number of keys in a hash table, based on a tuple of (long_integer, integer). The standard python dictionary works well for small numbers of keys, but starts to perform badly for me inserting roughly 5M keys:
# keys dictionary metakit (both using psyco) ------ ---------- ------- 1M 8.8s 22.2s 2M 24.0s 43.7s 5M 115.3s 105.4s Does anyone know of a fast hash module which is more optimal for large datasets ? p.s. Disk-based DBs are out of the question because most key lookups will result in a miss, and lookup time is critical for this application. Cheers, -- Chris Foote <[EMAIL PROTECTED]> Inetd Pty Ltd T/A HostExpress Web: http://www.hostexpress.com.au Blog: http://www.hostexpress.com.au/drupal/chris Phone: (08) 8410 4566 _______________________________________________ sapug mailing list [email protected] http://mail.python.org/mailman/listinfo/sapug
