You could use multiple tables to get around the file size limit.

Decide how many rows would go in each table and then use some mechanism for 
assigning a unique ID to each row. You would then be able to determine which 
table a specific row is in with an integer division.

table # = <global row #> / <rows per table>

local row # = <global row #> % <rows per table>

The Telemetry Data Warehouse for the Hubble Space Telescope divides the data up 
by time - each data table contains one calendar year's worth of telemetry data.

Just a couple of ideas.

Donald

 -------------- Original message ----------------------
From: Suresh Thalamati <[EMAIL PROTECTED]>
> redcloud wrote:
> > Hi! I need to build a SQL table containing 1000000000 (!!!) rows. But i 
> > filled up a table with 20000000 rows (file size of table 4GB) and my 
> > filesystem denied to go on filling up the table. My question is: can 
> > derby db build "infinitive" size table by chunking in multiple files?
> > 
> 
> No. Currently a table maps to a single file in Derby. Table size is 
> limited by the size of the file that can be created on a file system.

Reply via email to