You could use multiple tables to get around the file size limit. Decide how many rows would go in each table and then use some mechanism for assigning a unique ID to each row. You would then be able to determine which table a specific row is in with an integer division.
table # = <global row #> / <rows per table> local row # = <global row #> % <rows per table> The Telemetry Data Warehouse for the Hubble Space Telescope divides the data up by time - each data table contains one calendar year's worth of telemetry data. Just a couple of ideas. Donald -------------- Original message ---------------------- From: Suresh Thalamati <[EMAIL PROTECTED]> > redcloud wrote: > > Hi! I need to build a SQL table containing 1000000000 (!!!) rows. But i > > filled up a table with 20000000 rows (file size of table 4GB) and my > > filesystem denied to go on filling up the table. My question is: can > > derby db build "infinitive" size table by chunking in multiple files? > > > > No. Currently a table maps to a single file in Derby. Table size is > limited by the size of the file that can be created on a file system.
