at 12:33 AM
To: Cassandra User user@cassandra.apache.org
Subject: Re: Exactly one wide row per node for a given CF?
But this becomes troublesome if I add or remove nodes. What effectively I
want is to partition on the unique id of the record modulus N (id % N;
where N is the number of nodes
, 2013 at 12:33 AM
To: Cassandra User user@cassandra.apache.org
Subject: Re: Exactly one wide row per node for a given CF?
But this becomes troublesome if I add or remove nodes. What effectively I
want is to partition on the unique id of the record modulus N (id % N; where
N is the number
comments below
On Dec 9, 2013, at 11:33 PM, Aaron Morton aa...@thelastpickle.com wrote:
But this becomes troublesome if I add or remove nodes. What effectively I
want is to partition on the unique id of the record modulus N (id % N; where
N is the number of nodes).
This is exactly the
...@thelastpickle.com
Reply-To: user@cassandra.apache.org
Date: Tuesday, December 10, 2013 at 12:33 AM
To: Cassandra User user@cassandra.apache.org
Subject: Re: Exactly one wide row per node for a given CF?
But this becomes troublesome if I add or remove nodes. What effectively I
want is to partition
boundaries.
Any suggestions on a better pattern?
Thanks
Robert
From: Aaron Morton aa...@thelastpickle.com
Reply-To: user@cassandra.apache.org
Date: Tuesday, December 10, 2013 at 12:33 AM
To: Cassandra User user@cassandra.apache.org
Subject: Re: Exactly one wide row per node for a given CF
But this becomes troublesome if I add or remove nodes. What effectively I
want is to partition on the unique id of the record modulus N (id % N; where
N is the number of nodes).
This is exactly the problem consistent hashing (used by cassandra) is designed
to solve. If you hash the key and
Basically this desire all stems from wanting efficient use of memory.
Do you have any real latency numbers you are trying to tune ?
Otherwise this sounds a little like premature optimisation.
Cheers
-
Aaron Morton
New Zealand
@aaronmorton
Co-Founder Principal Consultant
Pretty much yes. Although I think it’d be nice if Cassandra handled such a
case, I’ve resigned to the fact that it cannot at the moment. The workaround
will be to partition on the LSB portion of the id (giving 256 rows spread
amongst my nodes) which allows room for scaling, and then cluster
So Basically you want to create a cluster of multiple unique keys, but data
which belongs to one unique should be colocated. correct?
-Vivek
On Tue, Dec 3, 2013 at 10:39 AM, onlinespending onlinespend...@gmail.comwrote:
Subject says it all. I want to be able to randomly distribute a large set