Hi All,
When I run ".dump" on a small table I created at the prompt, it works
fine. But when I try to run .dump on a big table it only gives me the
table schema, pauses, then stops (see below for example). It happens
when I run in batch from the command line, too. I am able to select
from the large table just fine ('select count(*) from nlsy79;' gives
25374, group by's work fine, etc.).
I am working with a modified 3.6.7 codebase (edited various max_*
modified to be very large). Though I can't imagine why that would
matter, unless there is something that won't dump unless under a
certain max.
(Note that I am not looking for solutions that require that I truncate
the table or get rid of columns, because of the nature of the
problem.)
Example 1, first table, works fine:
sqlite> .dump blah
BEGIN TRANSACTION;
CREATE TABLE blah (i int, s text);
INSERT INTO "blah" VALUES(10,'blah blah');
COMMIT;
Example 2, second table, notice lack of insert statements:
sqlite> .dump NLSY79
BEGIN TRANSACTION;
CREATE TABLE NLSY79 ( "Obs" integer
, "R0000100" integer
, "R0000149" integer
, "R0000150" integer
--... (SNIP some 2000+ columns)
, "W0072300" integer
, "W0072400" integer
);
COMMIT;
Thanks for all the help!
_______________________________________________
sqlite-users mailing list
[email protected]
http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users