Your RF is only 1, so the data only exists on one node. This is not typically how Cassandra is used. If you need the high availability and low latency, you typically set RF to 3 per DC.
How many event_datetime records can you have per pkey? How many pkeys (roughly) do you have? In general, you only want to have at most 100 MB of data per partition (pkey). If it is larger than that, I would expect some timeouts. And because only one node has the data, a single timeout means you won’t get any data. Server timeouts default to just 10 seconds. The secret to Cassandra is to always select your data by at least the primary key (which you are doing). So, I suspect you either have very wide rows or lots of tombstones. Since you mention lots of deletes, I am thinking it could be tombstones. Are you getting any tombstone warnings or errors in your system.log? When you delete, are you deleting a full partition? If you are deleting just part of a partition over and over, I think you will be creating too many tombstones. I try to design my data partitions so that deletes are for a full partition. Then I won’t be reading through 1000s (or more) tombstones trying to find the live data. Sean Durity From: Marco Gasparini <marco.gaspar...@competitoor.com> Sent: Thursday, December 27, 2018 3:01 AM To: user@cassandra.apache.org Subject: Re: [EXTERNAL] Writes and Reads with high latency Hello Sean, here my schema and RF: ------------------------------------------------------------------------- CREATE KEYSPACE my_keyspace WITH replication = {'class': 'NetworkTopologyStrategy', 'DC1': '1'} AND durable_writes = true; CREATE TABLE my_keyspace.my_table ( pkey text, event_datetime timestamp, agent text, ft text, ftt text, some_id bigint, PRIMARY KEY (pkey, event_datetime) ) WITH CLUSTERING ORDER BY (event_datetime DESC) AND bloom_filter_fp_chance = 0.01 AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'} AND comment = '' AND compaction = {'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy', 'max_threshold': '32', 'min_threshold': '4'} AND compression = {'chunk_length_in_kb': '64', 'class': 'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND dclocal_read_repair_chance = 0.1 AND default_time_to_live = 0 AND gc_grace_seconds = 90000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair_chance = 0.0 AND speculative_retry = '99PERCENTILE'; ------------------------------------------------------------------------- Queries I make are very simple: select pkey, event_datetime, ft, some_id, ftt from my_keyspace.my_table where pkey = ? limit ?; and insert into my_keyspace.my_table (event_datetime, pkey, agent, some_id, ft, ftt) values (?,?,?,?,?,?); About Retry policy, the answer is yes, actually when a write fails I store it somewhere else and, after a period, a try to write it to Cassandra again. This way I can store almost all my data, but when the problem is the read I don't apply any Retry policy (but this is my problem) Thanks Marco Il giorno ven 21 dic 2018 alle ore 17:18 Durity, Sean R <sean_r_dur...@homedepot.com<mailto:sean_r_dur...@homedepot.com>> ha scritto: Can you provide the schema and the queries? What is the RF of the keyspace for the data? Are you using any Retry policy on your Cluster object? Sean Durity From: Marco Gasparini <marco.gaspar...@competitoor.com<mailto:marco.gaspar...@competitoor.com>> Sent: Friday, December 21, 2018 10:45 AM To: user@cassandra.apache.org<mailto:user@cassandra.apache.org> Subject: [EXTERNAL] Writes and Reads with high latency hello all, I have 1 DC of 3 nodes in which is running Cassandra 3.11.3 with consistency level ONE and Java 1.8.0_191. Every day, there are many nodejs programs that send data to the cassandra's cluster via NodeJs cassandra-driver. Every day I got like 600k requests. Each request makes the server to: 1_ READ some data in Cassandra (by an id, usually I get 3 records), 2_ DELETE one of those records 3_ WRITE the data into Cassandra. So every day I make many deletes. Every day I find errors like: "All host(s) tried for query failed. First host tried, 10.8.0.10:9042<https://urldefense.proofpoint.com/v2/url?u=http-3A__10.8.0.10-3A9042_&d=DwMFaQ&c=MtgQEAMQGqekjTjiAhkudQ&r=aC_gxC6z_4f9GLlbWiKzHm1vucZTtVYWDDvyLkh8IaQ&m=Y2zNzOyvqOiHqZ5yvB1rO_X6C-HivNjXYN0bLLL-yZQ&s=2v42cyvuxcXJ0oMfUrRcY-kRno1SkM4CTEMi4n1k0Wo&e=>: Host considered as DOWN. See innerErrors...." "Server timeout during write query at consistency LOCAL_ONE (0 peer(s) acknowledged the write over 1 required)...." "Server timeout during write query at consistency SERIAL (0 peer(s) acknowledged the write over 1 required)...." "Server timeout during read query at consistency LOCAL_ONE (0 peer(s) acknowledged the read over 1 required)...." nodetool tablehistograms tells me this: Percentile SSTables Write Latency Read Latency Partition Size Cell Count (micros) (micros) (bytes) 50% 8.00 379.02 1955.67 379022 8 75% 10.00 785.94 155469.30 654949 17 95% 12.00 17436.92 268650.95 1629722 35 98% 12.00 25109.16 322381.14 2346799 42 99% 12.00 30130.99 386857.37 3379391 50 Min 0.00 6.87 88.15 104 0 Max 12.00 43388.63 386857.37 20924300 179 in the 99% I noted that write and read latency is pretty high, but I don't know how to improve that. I can provide more statistics if needed. Is there any improvement I can make to the Cassandra's configuration in order to not to lose any data? Thanks Regards Marco ________________________________ The information in this Internet Email is confidential and may be legally privileged. It is intended solely for the addressee. Access to this Email by anyone else is unauthorized. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is prohibited and may be unlawful. When addressed to our clients any opinions or advice contained in this Email are subject to the terms and conditions expressed in any applicable governing The Home Depot terms of business or client engagement letter. The Home Depot disclaims all responsibility and liability for the accuracy and content of this attachment and for any damages or losses arising from any inaccuracies, errors, viruses, e.g., worms, trojan horses, etc., or other items of a destructive nature, which may be contained in this attachment and shall not be liable for direct, indirect, consequential or special damages in connection with this e-mail message or its attachment. ________________________________ The information in this Internet Email is confidential and may be legally privileged. It is intended solely for the addressee. Access to this Email by anyone else is unauthorized. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is prohibited and may be unlawful. When addressed to our clients any opinions or advice contained in this Email are subject to the terms and conditions expressed in any applicable governing The Home Depot terms of business or client engagement letter. The Home Depot disclaims all responsibility and liability for the accuracy and content of this attachment and for any damages or losses arising from any inaccuracies, errors, viruses, e.g., worms, trojan horses, etc., or other items of a destructive nature, which may be contained in this attachment and shall not be liable for direct, indirect, consequential or special damages in connection with this e-mail message or its attachment.