Okay, I see that you can tell that it's running by doing a GET on the
database in question and looking for the "compact_running":true .  However,
I don't seem to see any changes in the db's file size.

On Mon, May 14, 2012 at 3:42 PM, Tim Tisdall <[email protected]> wrote:

> Yes, I did it with a PUT for each id.  When you call for compaction, is
> there a way to see the progress or a way to know if it's done?
>
>
> On Mon, May 14, 2012 at 3:20 PM, Paul Davis 
> <[email protected]>wrote:
>
>> How did you insert them? If you did a PUT per docid you'll still want
>> to compact afterwards.
>>
>> On Mon, May 14, 2012 at 2:13 PM, Tim Tisdall <[email protected]> wrote:
>> > I've got several gigabytes of data that I'm trying to store in a
>> couchdb on
>> > a single machine.  I've placed a section of the data in an sqlite db and
>> > the file is about 5.9gb.  I'm currently placing the same data into
>> couchdb
>> > and while it hasn't finished yet, the file size is already 10gb and
>> > continuing to grow.  The sqlite database is essentially a table of ids
>> with
>> > a json block of text for each, so I figured the couchdb wouldn't be too
>> > much different in size.
>> >
>> > Does anyone have some recommendations on how to reduce the size of the
>> db?
>> >  Right now I've only inserted data and have not made any "updates" to
>> > documents, so there should be no revision copies to be cleared away.
>>
>
>

Reply via email to