Thanks Jeff for the information.
On Wed, 10 Nov 2021 at 21:08, Jeff Jirsa wrote:
> This type of delete - which doesnt supply a user_id, so it's deleting a
> range of rows - creates what is known as a range tombstone. It's not tied
> to any given cell, as it covers a range of cells, and
This type of delete - which doesnt supply a user_id, so it's deleting a
range of rows - creates what is known as a range tombstone. It's not tied
to any given cell, as it covers a range of cells, and supersedes/shadows
them when merged (either in the read path or compaction path).
On Wed, Nov
HI Team,
I have one table below and want to delete data on this table.
DELETE FROM game.tournament USING TIMESTAMP 161692578000 WHERE
tournament_id = 1 AND version_id = 1 AND partition_id = 1;
Cassandra internally manages the timestamp of each column when some data is
updated on the
Glad you got it figured out, but I'm confused about the @Version
annotation. The DataStax Java Driver just handles statements, as far as I
know it's never going to modify statement text. It sounds like you're
using an entity mapping framework on top of the java driver, which uses
@Version for
Hi All.
I think we got the root-cause.
One of the fields in one of the class was marked with "@Version"
annotation, which was causing the Cassandra-Java-Driver to insert "If Not
Exists" in the insert query, thus invoking SERIAL consistency-level.
We removed the annotation (didn't really need
What Eric means is that SERIAL consistency is a special type of consistency
that is only invoked for a subset of operations: those that use
CAS/lightweight transactions, for example "IF NOT EXISTS" queries.
The differences between CAS operations and standard operations are
significant and there
Hmm... ok.
Ideally, we require ::
a)
The intra-DC-node-syncing takes place at the statement/query level.
b)
The inter-DC-node-syncing takes place at cassandra level.
That way, we don't spend too much delay at the statement/query level.
For the so-called CAS/lightweight transactions, the
Hi All.
I have a 2*2 Network-Topology Replication setup, and I run my application
via DataStax-driver.
I frequently get the errors of type ::
*Cassandra timeout during write query at consistency SERIAL (3 replica were
required but only 0 acknowledged the write)*
I have already tried passing a
Serial consistency gets invoked at the protocol level when doing
lightweight transactions such as CAS operations. If you're expecting that
your topology is RF=2, N=2, it seems like some keyspace has RF=3, and so
there aren't enough nodes available to satisfy serial consistency.
See
Hi Eric,
I am sorry, but I don't understand.
If there had been some issue in the configuration, then the
consistency-issue would be seen everytime (I guess).
As of now, the error is seen sometimes (probably 30% of times).
On Mon, Nov 2, 2015 at 10:24 PM, Eric Stevens wrote:
of time and now I am
modeling a database for study purposes. During my modeling I have faced a
doubt, what is a good partition key? Is partition key direct related with
my query performance? What is the best practices?
Just to study case, let's suppose I have a column family where is
inserted
Hello folks
I am studying Cassandra for a short a period of time and now I am modeling
a database for study purposes. During my modeling I have faced a doubt,
what is a good partition key? Is partition key direct related with my query
performance? What is the best practices?
Just to study case
guilherme@gmail.com wrote:
Hello folks
I am studying Cassandra for a short a period of time and now I am modeling
a database for study purposes. During my modeling I have faced a doubt,
what is a good partition key? Is partition key direct related with my query
performance? What is the best
Generally Ive seen it recommended to do a composite CF since it gives you more
flexibility and its easier to debug. You can get some performance improvements
by storing a serialized blob (a lot of data you can represent much smaller this
way by factor of 10 or more if clever) to represent your
Dear All,
We have a requirement to store 'N' columns of an entity in a CF. Mostly this is
write once and read many times. What is the best way to store the data?
Composite CF
Simple CF with value as protobuf extracted data
Both provides extendable columns which is a requirement for our usage.
Dear all
I have stored my data into Cassandra database in the format tickerID_date.
Now when I specify the row key range like 1_2012/05/24(start) to
1_2012/05/27(end) it says that the end key md5 value is lesser than start key
md5 value. So I changed my start key to 1_2012/05/27 and end key
-Original Message-
From: Prakrati Agrawal prakrati.agra...@mu-sigma.com
Date: Mon, 28 May 2012 04:39:46
To: user@cassandra.apache.orguser@cassandra.apache.org
Reply-To: user@cassandra.apache.org
Subject: Doubt in Row key range scan
Dear all
I have stored my data into Cassandra database
You are using the Random Partitioner.
Using the RP is a good thing because you avoid hot spots, but it has
its defaults too. You can't scan a slice of row, they won't be ordered
because all your keys are stored using their md5 values.
You should review your data model to use columns to order
Please could you tell me how to tweak my data model to rely on 2nd index ?
Thank you
Prakrati Agrawal | Developer - Big Data(ID)| 9731648376 | www.mu-sigma.com
From: Pierre Chalamet [mailto:pie...@chalamet.net]
Sent: Monday, May 28, 2012 3:31 PM
To: user@cassandra.apache.org
Subject: Re: Doubt
Data(ID)| 9731648376 | www.mu-sigma.com
From: Pierre Chalamet [mailto:pie...@chalamet.net]
Sent: Monday, May 28, 2012 3:31 PM
To: user@cassandra.apache.org
Subject: Re: Doubt in Row key range scan
Hi,
It's normal.
Keys to replicas are determined with a hash (md5) when using
On Wednesday 22 of February 2012, Rishabh Agrawal wrote:
I have installed CQL drivers for python. When I try execute cqlsh I get
following error
cql-1.0.3$ cqlsh localhost 9160
(...)
File /usr/local/lib/python2.7/dist-packages/cql/cassandra/ttypes.py,
line 7, in module from thrift.Thrift
Thanks for the reply
I installed 0.8.0 drift package. But still problem persists.
-Original Message-
From: Mateusz Korniak [mailto:mateusz-li...@ant.gliwice.pl]
Sent: Wednesday, February 22, 2012 1:47 PM
To: user@cassandra.apache.org
Subject: Re: Doubt regarding CQL
On Wednesday 22
: Wednesday, February 22, 2012 1:47 PM
To: user@cassandra.apache.org
Subject: Re: Doubt regarding CQL
On Wednesday 22 of February 2012, Rishabh Agrawal wrote:
I have installed CQL drivers for python. When I try execute cqlsh I
get following error cql-1.0.3$ cqlsh localhost 9160
(...)
File
FYI .. I am using 1.0.7 version on Ubuntu 11.10
Need help asap
From: Rishabh Agrawal
Sent: Wednesday, February 22, 2012 11:49 AM
To: user@cassandra.apache.org
Subject: Doubt regarding CQL
Hello
I have installed CQL drivers for python. When I try execute cqlsh I get
following error
cql-1.0.3
...@yahoo.co.in wrote:
From: Thamizh tceg...@yahoo.co.in
Subject: Re: multi-node cassandra config doubt
To: user@cassandra.apache.org
Date: Thursday, 25 August, 2011, 9:01 PM
Hi Aaron,
Thanks a lot for your suggestions. I have got exhausted with below error. It
would great if you point me what went
morton aa...@thelastpickle.com wrote:
From: aaron morton aa...@thelastpickle.com
Subject: Re: multi-node cassandra config doubt
To: user@cassandra.apache.org
Date: Thursday, 25 August, 2011, 3:45 AM
Jump on the machine that raised the error and see if you can ssh to node01.
or try using ip address
cluster configuration doubt.
I have configured 3 nodes of cluster using Cassandra-0.8.4 and getting error
when I ran Map/Reduce job which uploads records from HDFS to Cassandra.
Here are my 3 nodes cluster config file (cassandra.yaml) for Cassandra:
node01:
seeds: node01,node02,node03
has been used across all the nodes.
Kindly help me to resolve this issue?
Regards,
Thamizhannal P
--- On Wed, 24/8/11, aaron morton aa...@thelastpickle.com wrote:
From: aaron morton aa...@thelastpickle.com
Subject: Re: multi-node cassandra config doubt
To: user@cassandra.apache.org
Date
Hi All,
This is regarding multi-node cluster configuration doubt.
I have configured 3 nodes of cluster using Cassandra-0.8.4 and getting error
when I ran Map/Reduce job which uploads records from HDFS to Cassandra.
Here are my 3 nodes cluster config file (cassandra.yaml) for Cassandra:
node01
29 matches
Mail list logo