Hi, I uploaded about 11K + docs total 230MB or so of data to a 0.9 instance on Ubuntu. Db name is 'plist'
curl http://localhost:5984/plist gives {"db_name":"plist","doc_count":11036,"doc_del_count":0,"update_seq":11036,"purge_seq":0, "compact_running":false,"disk_size":243325178,"instance_start_time":"1246228896723181"} suggesting a non-corrupt db curl http://localhost:5984/plist/_all_docs gives {"id":"adnanmoh","key":"adnanmoh","value":{"rev":"1-663736558"}}, {"id":"adnen.chockri","key":"adnen.chockri","value":{"rev":"1-1209124545"}}, curl: (56) Received problem 2 in the chunky parser <<--------- note curl error {"id":"ado.adamu","key":"ado.adamu","value":{"rev":"1-4226951654"}} suggesting a chunked data transfer error couchdb-lucene error message in couchdb.stderr reads [...] [couchdb-lucene] INFO Indexing plist from scratch. [couchdb-lucene] ERROR Error updating index. java.io.IOException: CRLF expected at end of chunk: 83/101 at org.apache.commons.httpclient.ChunkedInputStream.readCRLF(ChunkedInputStream.java:207) at org.apache.commons.httpclient.ChunkedInputStream.nextChunk(ChunkedInputStream.java:219) at org.apache.commons.httpclient.ChunkedInputStream.read(ChunkedInputStream.java:176) at org.apache.commons.httpclient.ChunkedInputStream.read(ChunkedInputStream.java:196) at org.apache.commons.httpclient.ChunkedInputStream.exhaustInputStream(ChunkedInputStream.java:369) at org.apache.commons.httpclient.ChunkedInputStream.close(ChunkedInputStream.java:346) at java.io.FilterInputStream.close(FilterInputStream.java:159) at org.apache.commons.httpclient.AutoCloseInputStream.notifyWatcher(AutoCloseInputStream.java:194) at org.apache.commons.httpclient.AutoCloseInputStream.close(AutoCloseInputStream.java:158) at com.github.rnewson.couchdb.lucene.Database.execute(Database.java:141) at com.github.rnewson.couchdb.lucene.Database.get(Database.java:107) at com.github.rnewson.couchdb.lucene.Database.getAllDocsBySeq(Database.java:82) at com.github.rnewson.couchdb.lucene.Index$Indexer.updateDatabase(Index.java:229) at com.github.rnewson.couchdb.lucene.Index$Indexer.updateIndex(Index.java:178) at com.github.rnewson.couchdb.lucene.Index$Indexer.run(Index.java:90) at java.lang.Thread.run(Thread.java:595) suggesting a chunking problem again. Who is creating this problem - my data? CouchDB chunking ? Help? 37% of all statistics are made up on the spot ------------------------------------------------------------------------------------- Nitin Borwankar [email protected]
