> I have a simple table with five columns and 450,000 rows.  In SQLiteSpy,
> I can run "SELECT * FROM trend_data" and get all 450,000 rows in 4.5
> seconds.  But in my program, if I use sqlite3_prepare() and
> sqlite3_step() until I run out of data, it takes 55 seconds to get
> through all rows. 

Selecting 100K rows/second without ORDER BY seems reasonable on a fast machine.

What timing do you get for this command?

  time sqlite3 your.db "SELECT * FROM trend_data" | wc -l

Run the command more than once, as the first timing is always slower.



      
____________________________________________________________________________________Fussy?
 Opinionated? Impossible to please? Perfect.  Join Yahoo!'s user panel and lay 
it on us. http://surveylink.yahoo.com/gmrs/yahoo_panel_invite.asp?a=7 


-----------------------------------------------------------------------------
To unsubscribe, send email to [EMAIL PROTECTED]
-----------------------------------------------------------------------------

Reply via email to