ADARSH KUMAR created CASSANDRA-11920:
----------------------------------------
Summary: Not able to set bloom_filter_fp_chance as .00001
Key: CASSANDRA-11920
URL: https://issues.apache.org/jira/browse/CASSANDRA-11920
Project: Cassandra
Issue Type: Bug
Components: Lifecycle, Local Write-Read Paths
Reporter: ADARSH KUMAR
Hi,
I was doing some bench-marking on bloom_filter_fp_chance values. Everything
worked fine for values .01(default for STCS), .001, .0001. But when I set
bloom_filter_fp_chance = .00001 i observed following behaviour:
1). Reads and writes looked normal from cqlsh.
2). SSttables are never created.
3). It just creates two files (*-Data.db and *-index.db) of size 0kb.
4). nodetool flush does not work and produce following exception:
java.lang.UnsupportedOperationException: Unable to satisfy 1.0E-5 with 20
buckets per element
at
org.apache.cassandra.utils.BloomCalculations.computeBloomSpec(BloomCalculations.java:150)
.....
I checked BloomCalculations class and following lines are responsible for this
exception:
if (maxFalsePosProb < probs[maxBucketsPerElement][maxK]) {
throw new UnsupportedOperationException(String.format("Unable to satisfy
%s with %s buckets per element",
maxFalsePosProb,
maxBucketsPerElement));
}
>From the code it looks like a hard coaded validation (unless we can change
>the nuber of buckets).
So, if this validation is hard coaded then why it is even allowed to set such
value of bloom_fileter_fp_chance, that can prevent ssTable generation?
Please correct this issue.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)