Oh and when I use Futon and try to browse the docs around where curl gives an error, when I hit the page containing the records around the error Futon just spins and doesn't render the page.
Data corruption? Nitin 37% of all statistics are made up on the spot ------------------------------------------------------------------------------------- Nitin Borwankar [email protected] On Mon, Jun 29, 2009 at 9:11 PM, Nitin Borwankar <[email protected]>wrote: > > Hi, > > I uploaded about 11K + docs total 230MB or so of data to a 0.9 instance on > Ubuntu. > Db name is 'plist' > > curl http://localhost:5984/plist gives > > > {"db_name":"plist","doc_count":11036,"doc_del_count":0,"update_seq":11036,"purge_seq":0, > > "compact_running":false,"disk_size":243325178,"instance_start_time":"1246228896723181"} > > suggesting a non-corrupt db > > curl http://localhost:5984/plist/_all_docs gives > > {"id":"adnanmoh","key":"adnanmoh","value":{"rev":"1-663736558"}}, > > {"id":"adnen.chockri","key":"adnen.chockri","value":{"rev":"1-1209124545"}}, > curl: (56) Received problem 2 in the chunky > parser <<--------- note curl error > {"id":"ado.adamu","key":"ado.adamu","value":{"rev":"1-4226951654"}} > > suggesting a chunked data transfer error > > > couchdb-lucene error message in couchdb.stderr reads > > [...] > > [couchdb-lucene] INFO Indexing plist from scratch. > [couchdb-lucene] ERROR Error updating index. > java.io.IOException: CRLF expected at end of chunk: 83/101 > at > org.apache.commons.httpclient.ChunkedInputStream.readCRLF(ChunkedInputStream.java:207) > at > org.apache.commons.httpclient.ChunkedInputStream.nextChunk(ChunkedInputStream.java:219) > at > org.apache.commons.httpclient.ChunkedInputStream.read(ChunkedInputStream.java:176) > at > org.apache.commons.httpclient.ChunkedInputStream.read(ChunkedInputStream.java:196) > at > org.apache.commons.httpclient.ChunkedInputStream.exhaustInputStream(ChunkedInputStream.java:369) > at > org.apache.commons.httpclient.ChunkedInputStream.close(ChunkedInputStream.java:346) > at java.io.FilterInputStream.close(FilterInputStream.java:159) > at > org.apache.commons.httpclient.AutoCloseInputStream.notifyWatcher(AutoCloseInputStream.java:194) > at > org.apache.commons.httpclient.AutoCloseInputStream.close(AutoCloseInputStream.java:158) > at > com.github.rnewson.couchdb.lucene.Database.execute(Database.java:141) > at com.github.rnewson.couchdb.lucene.Database.get(Database.java:107) > at > com.github.rnewson.couchdb.lucene.Database.getAllDocsBySeq(Database.java:82) > at > com.github.rnewson.couchdb.lucene.Index$Indexer.updateDatabase(Index.java:229) > at > com.github.rnewson.couchdb.lucene.Index$Indexer.updateIndex(Index.java:178) > at com.github.rnewson.couchdb.lucene.Index$Indexer.run(Index.java:90) > at java.lang.Thread.run(Thread.java:595) > > > suggesting a chunking problem again. > > Who is creating this problem - my data? CouchDB chunking ? > > Help? > > > > 37% of all statistics are made up on the spot > > ------------------------------------------------------------------------------------- > Nitin Borwankar > [email protected] >
