Hi Radim
Unfortunately number of compaction tasks is not overestimated. The number
is decremented one-by-one and this process takes several hours for our 40GB
node(( Also, when a lot of compaction tasks appears, we see that total disk
space used (via JMX) is doubled and Cassandra really tries to
Thanks for the reply and sorry for being bull - headed.
Once you're past the stage where you've decided its distributed, and NoSQL and
cassandra out of all the NoSQL options,Now to count something, you can do it in
different ways in cassandra. In all the ways you want to use cassandra's best
From my point of view an other problem with using the standard column
family for counting is transactions. Cassandra lacks of them, so if you're
multithreaded updating counters, how will you keep track of that? Yes, I'm
aware of software like Zookeeper to do that, however I'm not sure whether
See my comments inline
2012/9/25 Aaron Turner synfina...@gmail.com
On Mon, Sep 24, 2012 at 10:02 AM, Віталій Тимчишин tiv...@gmail.com
wrote:
Why so?
What are pluses and minuses?
As for me, I am looking for number of files in directory.
700GB/512MB*5(files per SST) = 7000 files, that
hi,
One node in my cluster stay in joining. I found a jira about this, which is
fixed,but still sees the similar thing. This is a node I remove the token first
becauseit did not boot correctly and re-joined in the cluster without any
pre-set token(shouldI set the previous token?).
As you see
I've recently noticed several threads about Cassandra
Counters inconsistencies and started seriously think about possible
workarounds like store realtime counters in Redis and dump them daily to
Cassandra.
So general question, should I rely on Counters if I want 100% accuracy?
Thanks, Ed
On Tue,
Thanks a lot for helping. We came to the same decision clustering one
report to multiple cassandra rows (sorted buckets of report rows) and
manage clusters on client side.
On Tue, Sep 25, 2012 at 5:28 AM, aaron morton aa...@thelastpickle.com wrote:
What exactly is the problem with big rows?
Is this a bug? I'm using Cassandra 1.0.11:
INFO 13:45:43,750 Compacting
[SSTableReader(path='d:\data\Traxis\Parameters-hd-47-Data.db'),
SSTableReader(path='d:\data\Traxis\Parameters-hd-44-Data.db'),
SSTableReader(path='d:\data\Traxis\Parameters-hd-46-Data.db'),
@Edward,
We use counters in production with Cassandra 1.0.5. Though since our
application is sensitive to write latency and we are seeing problems with
Frequent Young Garbage Collections, and also we just do increments
(decrements have caused problems for some people)
We don't see inconsistencies
If you need anything added/fixed, just let PlayOrm know. PlayOrm has been able
to quickly add so far…that may change as more and more requests come but so far
PlayOrm seems to have managed to keep up.
We are using it live by the way already. It works out very well so far for us
(We have 5000
So general question, should I rely on Counters if I want 100% accuracy?
No.
Even not considering potential bugs, counters being not idempotent, if you
get a TimeoutException during a write (which can happen even in relatively
normal conditions), you won't know if the increment went in or not
@Sylvain
In a relatively untroubled cluster, even timed out writes go through,
provided no messages are dropped. Which you can monitor on cassandra
nodes. We have 100% consistency on our production servers as we don't
see messages being dropped on our servers.
Though as you mention, there would
@Sylvain and @Rohit: Thanks for your answers.
On Tue, Sep 25, 2012 at 2:27 PM, Sylvain Lebresne sylv...@datastax.comwrote:
So general question, should I rely on Counters if I want 100% accuracy?
No.
Even not considering potential bugs, counters being not idempotent, if
you get a
In a relatively untroubled cluster, even timed out writes go through,
provided no messages are dropped.
This all depends of your definition of untroubled cluster, but to be
clear, in a cluster where a node dies (which for Cassandra is not
considered abnormal and will happen to everyone no
Dean,
In the playOrm data modeling, if I understood it correctly, every CF
has its own id, right? For instance, User would have its own ID, Activities
would have its own id, etc. What if I have a trillion activities? Wouldn't
be a problem to have 1 row id for each activity?
Cassandra
Just fyi that some of these are cassandra questions…
Dean,
In the playOrm data modeling, if I understood it correctly, every CF has
its own id, right?
No, each entity has a field annotated with @NoSqlId. That tells playOrm this
is the row key. Each INSTANCE of the entity is a row in
Hey guys,
I've begun to notice that read operations take a performance nose-dive
after a standard (full) repair of a fairly large column family: ~11 million
records. Interestingly, I've then noticed that read performance returns to
normal after a full scrub of the column family. Is it possible
There are settings in cassandra.yaml that will _gradually_ reduce the
available cache to zero if you are under constant memory pressure:
# Set to 1.0 to disable. snip
reduce_cache_sizes_at: *
reduce_cache_capacity_to: *
My experience is that the cache size will not return to the configured
Oh, and if you really want to scale very easily, just use play framework
1.2.5 ;). We use that and since it is stateless, to scale up, you simple
add more servers. Also, it's like coding in php or ruby, etc. etc as far
as development speed(no server restarts) so it's a pretty nice framework.
We
On Tue, Sep 25, 2012 at 10:36 AM, Віталій Тимчишин tiv...@gmail.com wrote:
See my comments inline
2012/9/25 Aaron Turner synfina...@gmail.com
On Mon, Sep 24, 2012 at 10:02 AM, Віталій Тимчишин tiv...@gmail.com
wrote:
Why so?
What are pluses and minuses?
As for me, I am looking for
On Tue, Sep 25, 2012 at 1:23 PM, puneet loya puneetl...@gmail.com wrote:
http://goo.gl/JcMcr
--
Eric Evans
Acunu | http://www.acunu.com | @acunu
This is cassandra 1.1.4
Describe shows DecimalType and I test setting comparator TO the DecimalType
and it fails (Realize I have never touched this column family until now except
for posting data which succeeded)
[default@unknown] use databus;
Authenticated to keyspace: databus
Hmmm, is rowkey validation asynchronous to the actually sending of the
data to cassandra?
I seem to be able to put an invalid type and GET that invalid data back
just fine even though my type was an int and the comparator was Decimal
BUT then in the logs I see a validation fail exception but I
As per our understanding we expect that when we move token then that node
will first sync up the data as per the new assigned token only after that
it will receive the requests for new range.
When you use nodetool move the node will receive write requests for the new
range. As well as
Hi there,
Is there a way to embed/package Cassandra with an other Java application
and maintain control over it? Is this done before? Are there any best
practices?
Why I want to do this? We want to offer as less as configuration as
possible to our customers, but only if it's possible without
On Wed, Jul 18, 2012 at 10:27 AM, Douglas Muth doug.m...@gmail.com wrote:
Even though keyspace test1 had a replication_factor of 1 to start
with, each of the above UPDATE KEYSPACE commands caused a new UUID to
be generated for the schema, which I assume is normal and expected.
I believe the
On Sun, Sep 23, 2012 at 12:24 PM, Aaron Turner synfina...@gmail.com wrote:
Leveled compaction've tamed space for us. Note that you should set
sstable_size_in_mb to reasonably high value (it is 512 for us with ~700GB
per node) to prevent creating a lot of small files.
512MB per sstable? Wow,
Check the logs on nodes 2 and 3 to see if the scrub started. The logs on 1
will be a good help with that.
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 24/09/2012, at 10:31 PM, Tamar Fraenkel ta...@tok-media.com wrote:
Hi!
I ran
No.
Versions are capable of reading previous file formats, but can only create
files in the current format.
File formats are listed here
https://github.com/apache/cassandra/blob/trunk/src/java/org/apache/cassandra/io/sstable/Descriptor.java#L52
Looking for a way to make this work.
I'd
I wonder now if get_range_slices call will ever look for data in row
cache. I don't see it in the codebase. Only the get call will check row
cache?
On Wed, Sep 26, 2012 at 12:11 AM, Charles Brophy cbro...@zulily.com wrote:
There are settings in cassandra.yaml that will _gradually_ reduce the
The are thrift connection threads.
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 25/09/2012, at 1:32 AM, rohit bhatia rohit2...@gmail.com wrote:
Hi
What are pool-2-thread-* threads. Someone mentioned Client
Connection Threads.
Can you provide some information on the queries and the size of the data they
traversed ?
The default maximum size for a single thrift message is 16MB, was it larger
than that ?
https://github.com/apache/cassandra/blob/trunk/conf/cassandra.yaml#L375
Cheers
-
Aaron Morton
Which one will be faster to insert?
In general Composite types have the same performance; the extra work is
insignificant.
(Assuming you don't create a type with 100 components.)
And which one will be faster to read by incremental id?
If you have to specify the full key to get a row by row
Nothing jumps out. Are you able to reproduce the fault on a test node ?
There were some schema change problems in the early 1.1X releases. Did you
enable compression via a schema change ?
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On
The DEFAULT_CACHING_STRATEGY is Caching.KEYS_ONLY but even configuring row
cache size to be greater zero
won't enable row cache. Why?
On Wed, Sep 26, 2012 at 9:44 AM, Manu Zhang owenzhang1...@gmail.com wrote:
I wonder now if get_range_slices call will ever look for data in row
cache. I don't
Hi All,
I have a 4 node cluster setup in 2 zones with NetworkTopology strategy and
strategy options for writing a copy to each zone, so the effective load on
each machine is 50%.
Symptom:
I have a column family that has gc grace seconds of 10 days (the default).
On 17th there was an insert done
36 matches
Mail list logo