For the example provided by you , are you saying you are getting two rows
for same pk1,pk2,time?
It may be a problem with your inserts when you are inserting multiple
distinct rows or to validate all nodes are in sync try fetching using
CONSISTENCY ALL in cql.
On 18-Aug-2017 9:37 PM, "Nathan McL
You can SELECT DISTINCT in CQL, however I would recommend against such a
pattern as it is very unlikely to be efficient, and prone to errors. A
distinct query will search every partition for the first live cell, which
could be buried behind a lot of tombstones. It's safe to say at some point
you wi
You need to run an incremental repair for sstables to be marked repaired.
However only if all of the data in that Sstable is repaired during the
repair will you end up with it being marked repaired, otherwise an
anticompaction will occur and split the unrepaired data into its own
sstable.
It's pret
Cassandra version 3.9, -unit version 3.1.3.2.
In my (first ever) unit test, I've coded:
@BeforeClass public static void initFakeCassandra() throws
InterruptedException, IOException, TTransportException
{
EmbeddedCassandraServerHelper.startEmbeddedCassandra( 2L );
}
Execution crashes d
There seem to be a lot of SSTables in a repaired state and a lot in an
unrepaired state. For example, for this one table, the logs report
TRACE [main] 2017-08-15 23:50:30,732 LeveledManifest.java:473 - L0 contains 2
SSTables (176997267 bytes) in Manifest@1217144872
TRACE [main] 2017-08-15 23:50:
I have a table that uses LeveledCompactionStrategy on Cassandra 2.2. At the
moment, it has two SSTables, both in level 1, one that's repaired and one that
isn't.
$ sstablemetadata lb-135366-big-Data.db | head
SSTable: /home/cassandra/data/my_keyspace/my_table/lb-135366-big
Partitioner: org.apache
Hi,
I use cassandra-count (github
https://github.com/brianmhess/cassandra-count) to count records in a
table, but I have wrong results.
When I export data with cqlsh /copy to csv, I have 1M records in my test
table, when I use cassandra-count I have different results for each node :
build/ca
Thanks Jeff I have a horrible feeling it may be. I'll get the errors from the
dev guys Monday and email the group, hopefully I can tie this down.
Thanks
Sent from my iPhone
> On 18 Aug 2017, at 17:24, Jeff Jirsa wrote:
>
> Cassandra-13004
___
Regrettably, this may be a manifestation of Cassandra-13004, which could
corrupt data being read at the time you issued the ALTER TABLE command
What type of issues are you seeing? Anything in the logs?
--
Jeff Jirsa
> On Aug 18, 2017, at 8:41 AM, Jonathan Baynes
> wrote:
>
> Hi
>
> Is t
@Sagar,
A query to get the data looks like this (primary key values included in the
query).
SELECT * FROM table WHERE pk1='2269202-onstreet_high' AND pk2=2017 AND
time='2017-07-18 03:15:00+';
(in actual practice, the queries in our code would use query a range of
time values).
@Cristophe
I
Hi
Is there anything I need to do after dropping a column and adding in a column
to flush cassandra of the changes? We are experiencing issues with our front
end application, and the developers are asking if the issue was caused by the
change in schema, as I've done the drop and add of a column
hi:
Is this sensor data, hence timestamp? Ho w are you generating this 'key'
field?Can you have only the 'key' field as primary key? Even if not, since
that field is a part of the PK may make such queries fast.
However, are there other attributes thst can be added that define unique
business key
Hi
what is the most efficient way to get a distinct key list from a big table
(aprox 20 mil inserts per minute) ?
equivalent to *select distinct key from my_table *for this table
*CREATE TABLE my_table (*
*key text,*
*timestamp bigint,*
*value double,*
*PRIMARY KEY (key, time
I am doing some on-the-job-learning on this newer feature of the 3.x line,
where the token generation algorithm will compensate for different size nodes
in a cluster. In fact, it is one of the main reasons I upgraded to 3.0.13,
because I have a number of original nodes in a cluster that are abou
I would preferably spin 2 JVMs inside the same hardware (if you double
everything) than having to deal with what Jeff stated.
Also certain operations are not really found of a large number of vnodes
(eg. repair). There was a lot of improvements in the 3.x release cycle, but
I do still tend to redu
15 matches
Mail list logo