Hello,

On Tue, Aug 16, 2011 at 5:37 PM, Randall Leeds <[email protected]> wrote:
> On Tue, Aug 16, 2011 at 17:03, Chris Stockton 
> <[email protected]>wrote:
>
> Since you are collecting and creating keys in the output object creating
> this single property made the output of reduce larger. CouchDB tries to
> detect reduce functions that don't actually reduce the data. If you know for
> sure that you are working with a bounded set of properties whose occurrences
> you would like to sum you may set reduce_limit=false in your configuration.
> The default is true so that users don't shoot themselves in the foot
> (especially because you cannot cancel a run-away reduce if you don't have
> access to the machine to kill the process).
>

Thanks Randall for your reply, I changed my view call to [1] and oddly
it still gives the same error, maybe I am doing something wrong? I
didn't see anywhere on couchdb wiki anything for reduce_limit.
Although I think long term that kind of scares me a little bit, if for
some reason we ran across some new data that caused a infinite reduce
due to a bug, our couchdbs would all get crippled, do I have any other
options here?

It would be great if I could impose a size limit for reduce, or even a
"minimum" size limit, as it is odd to trigger a reduce error on the
first record, making it have to run at least 100 times should be a
good test to see if the data is "shrinking" or at least remaining
constant. Not sure what to suggest here beyond that, I just think it
doesn't feel quite right, maybe someone has some better suggestion.

[1] http://<server>:59841/db_24/_design/test/_view/Grid?reduce_limit=false

Reply via email to