I was prototyping an app with Mongo and Couch to see which one I liked
the best (I've been flirting with couch for a long time, so I am
somewhat biased -- but, I am a flirt and mongo looked attractive).
Jan has a good article somewhere about why benchmarks are bad, so I
have nothing but the wall clock to back me up, but for our app, couch
was eating mongo on the writes, mongo ate couch for updates (upserts
are nice), and after finally understanding a little map/reduce we got
couch to do stuff that was very hard in mongo (we tried to use its
opslog to feed amqp and do map/reduce, but never got it working). YMMV
Troy
On Jan 28, 2010, at 6:06 AM, Sean Clark Hess wrote:
Oh, and my math was totally off before. The full 7 million rows
takes 1.7h
with MongoDB and 3h with Couch. (This is extrapolated from how long
it took
to insert the first 100k documents)
On Thu, Jan 28, 2010 at 7:02 AM, Troy Kruthoff <[email protected]>
wrote:
Just curious, what batch size did you use... I was just getting to
run
some test data to see where the sweet spot is for our hardware, I
remember
reading somewhere that someone thought it was around 3k docs.
Troy
On Jan 28, 2010, at 4:21 AM, Sean Clark Hess wrote:
Sweet... down to 28 minutes with bulk. Thanks
On Thu, Jan 28, 2010 at 4:25 AM, Sean Clark Hess
<[email protected]>
wrote:
Ah, I forgot about bulk! Thanks!
On Thu, Jan 28, 2010 at 4:24 AM, Alex Koshelev <[email protected]>
wrote:
How do you import data to CouchDB? Do you use _bulk API?
---
Alex Koshelev
On Thu, Jan 28, 2010 at 1:51 PM, Sean Clark Hess <[email protected]
>
wrote:
I'm trying to import 7 million rows into couch from an xml
document. If
I
use a database with a "normal" interface (comparing with Mongo
here),
the
process completes in 37 minutes. If I use couch, it takes 10
hours. I
think
it might be due to the overhead of the http interface, but I'm
not
sure.
Is there any way to get data in there faster?
~sean