In the mean time we opened traces and found that normal insert operations
does not have high latency. Delete operations have high latency. And all
deletes are range deletes. Is there any performance regression about range
deletes? Mean time of delete operations are 45 times higher than insert
operations.

On Thu, Jul 25, 2019, 11:46 AM mehmet bursali <bursal...@yahoo.com.invalid>
wrote:

> awesome! so we can make a further investigation by using cassandra
> exporter on this link.  https://github.com/criteo/cassandra_exporter
> This exporter gives detailed information for read/write operations on each
> column  by using metrics below..
>
>  org:apache:cassandra:metrics:columnfamily:.*  ( reads from table metrics in 
> cassandra 
> https://cassandra.apache.org/doc/latest/operating/metrics.html#table-metrics )
>
>
>
>
>
>
>
>
>
>
>
> On Wednesday, July 24, 2019, 11:51:28 PM GMT+3, CPC <acha...@gmail.com>
> wrote:
>
>
> Hi Mehmet,
>
> Yes prometheus and opscenter
>
> On Wed, 24 Jul 2019 at 17:09, mehmet bursali <bursal...@yahoo.com.invalid>
> wrote:
>
> hi,
> do you use any perfomance monitoring tool like prometheus?
>
>
>
>
> On Monday, July 22, 2019, 1:16:58 PM GMT+3, CPC <acha...@gmail.com>
> wrote:
>
>
> Hi everybody,
>
> State column contains "R" or "D" values. Just a single character. As Rajsekhar
> said, only difference is the table can contain high number of cell count.
> In the mean time we made a major compaction and data per node was 5-6 gb.
>
> On Mon, Jul 22, 2019, 10:56 AM Rajsekhar Mallick <raj.mallic...@gmail.com>
> wrote:
>
> Hello Team,
>
> The difference in write latencies between both the tables though
> significant,but the higher latency being 11.353 ms is still acceptable.
>
> Overall Writes not being an issue, but write latency for this particular
> table on the higher side does point towards data being written to the table.
> Few things which I noticed, is the data in cell count column in nodetool
> tablehistogram o/p for message_history_state table is scattered.
> The partition size histogram for the tables is consistent, but the column
> count histogram for the impacted table isn't uniform.
> May be we can start thinking on these lines.
>
> I would also wait for some expert advice here.
>
> Thanks
>
>
> On Mon,the 22 Jul, 2019, 12:31 PM Ben Slater, <ben.sla...@instaclustr.com>
> wrote:
>
> Is the size of the data in your “state” column variable? The higher write
> latencies at the 95%+ could line up with large volumes of data for
> particular rows in that column (the one column not in both tables)?
>
> Cheers
> Ben
>
> ---
>
>
> *Ben Slater**Chief Product Officer*
>
> <https://www.instaclustr.com/platform/>
>
> <https://www.facebook.com/instaclustr>   <https://twitter.com/instaclustr>
>    <https://www.linkedin.com/company/instaclustr>
>
> Read our latest technical blog posts here
> <https://www.instaclustr.com/blog/>.
>
> This email has been sent on behalf of Instaclustr Pty. Limited (Australia)
> and Instaclustr Inc (USA).
>
> This email and any attachments may contain confidential and legally
> privileged information.  If you are not the intended recipient, do not copy
> or disclose its content, but please reply to this email immediately and
> highlight the error to the sender and then immediately delete the message.
>
>
> On Mon, 22 Jul 2019 at 16:46, CPC <acha...@gmail.com> wrote:
>
> Hi guys,
>
> Any idea? I thought it might be a bug but could not find anything related
> on jira.
>
> On Fri, Jul 19, 2019, 12:45 PM CPC <acha...@gmail.com> wrote:
>
> Hi Rajsekhar,
>
> Here the details:
>
> 1)
>
> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY
> Total number of tables: 259
> ----------------
> Keyspace : tims
>         Read Count: 208256144
>         Read Latency: 7.655146714749506 ms
>         Write Count: 2218205275
>         Write Latency: 1.7826005103175133 ms
>         Pending Flushes: 0
>                 Table: MESSAGE_HISTORY
>                 SSTable count: 41
>                 Space used (live): 976964101899
>                 Space used (total): 976964101899
>                 Space used by snapshots (total): 3070598526780
>                 Off heap memory used (total): 185828820
>                 SSTable Compression Ratio: 0.8219217809913125
>                 Number of partitions (estimate): 8175715
>                 Memtable cell count: 73124
>                 Memtable data size: 26543733
>                 Memtable off heap memory used: 27829672
>                 Memtable switch count: 1607
>                 Local read count: 7871917
>                 Local read latency: 1.187 ms
>                 Local write count: 172220954
>                 Local write latency: 0.021 ms
>                 Pending flushes: 0
>                 Percent repaired: 0.0
>                 Bloom filter false positives: 130
>                 Bloom filter false ratio: 0.00000
>                 Bloom filter space used: 10898488
>                 Bloom filter off heap memory used: 10898160
>                 Index summary off heap memory used: 2480140
>                 Compression metadata off heap memory used: 144620848
>                 Compacted partition minimum bytes: 36
>                 Compacted partition maximum bytes: 557074610
>                 Compacted partition mean bytes: 155311
>                 Average live cells per slice (last five minutes): 
> 25.56639344262295
>                 Maximum live cells per slice (last five minutes): 5722
>                 Average tombstones per slice (last five minutes): 
> 1.8681948424068768
>                 Maximum tombstones per slice (last five minutes): 770
>                 Dropped Mutations: 97812
>
> ----------------
> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY_STATE
> Total number of tables: 259
> ----------------
> Keyspace : tims
>         Read Count: 208257486
>         Read Latency: 7.655137315414438 ms
>         Write Count: 2218218966
>         Write Latency: 1.7825896304427324 ms
>         Pending Flushes: 0
>                 Table: MESSAGE_HISTORY_STATE
>                 SSTable count: 5
>                 Space used (live): 6403033568
>                 Space used (total): 6403033568
>                 Space used by snapshots (total): 19086872706
>                 Off heap memory used (total): 6727565
>                 SSTable Compression Ratio: 0.271857664111622
>                 Number of partitions (estimate): 1396462
>                 Memtable cell count: 77450
>                 Memtable data size: 620776
>                 Memtable off heap memory used: 1338914
>                 Memtable switch count: 1616
>                 Local read count: 988278
>                 Local read latency: 0.518 ms
>                 Local write count: 109292691
>                 Local write latency: 11.353 ms
>                 Pending flushes: 0
>                 Percent repaired: 0.0
>                 Bloom filter false positives: 0
>                 Bloom filter false ratio: 0.00000
>                 Bloom filter space used: 1876208
>                 Bloom filter off heap memory used: 1876168
>                 Index summary off heap memory used: 410747
>                 Compression metadata off heap memory used: 3101736
>                 Compacted partition minimum bytes: 36
>                 Compacted partition maximum bytes: 129557750
>                 Compacted partition mean bytes: 17937
>                 Average live cells per slice (last five minutes): 
> 4.692893401015229
>                 Maximum live cells per slice (last five minutes): 258
>                 Average tombstones per slice (last five minutes): 1.0
>                 Maximum tombstones per slice (last five minutes): 1
>                 Dropped Mutations: 1344158
>
> 2)
>
> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims MESSAGE_HISTORY
> tims/MESSAGE_HISTORY histograms
> Percentile  SSTables     Write Latency      Read Latency    Partition Size    
>     Cell Count
>                               (micros)          (micros)           (bytes)
> 50%             3.00             20.50            454.83             14237    
>             17
> 75%            17.00             24.60           2346.80             88148    
>            103
> 95%            17.00             35.43          14530.76            454826    
>            924
> 98%            17.00             42.51          20924.30           1131752    
>           2299
> 99%            17.00             42.51          30130.99           1955666    
>           4768
> Min             0.00              3.97             73.46                36    
>              0
> Max            20.00            263.21          74975.55         386857368    
>         943127
>
> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims MESSAGE_HISTORY_STATE
> tims/MESSAGE_HISTORY_STATE histograms
> Percentile  SSTables     Write Latency      Read Latency    Partition Size    
>     Cell Count
>                               (micros)          (micros)           (bytes)
> 50%             5.00             20.50            315.85               924    
>              1
> 75%             6.00             35.43            379.02              5722    
>              7
> 95%            10.00           4055.27            785.94             61214    
>            310
> 98%            10.00          74975.55           3379.39            182785    
>            924
> 99%            10.00         107964.79          10090.81            315852    
>           1916
> Min             0.00              3.31             42.51                36    
>              0
> Max            10.00         322381.14          25109.16         129557750    
>        1629722
>
> 3) RF=3
>
> 4)CL  QUORUM
>
> 5) Single insert prepared statement. no LOGGED/UNLOGGED batch or LWT
>
>
> On Thu, 18 Jul 2019 at 20:51, Rajsekhar Mallick <raj.mallic...@gmail.com>
> wrote:
>
> Hello,
>
> Kindly post below details
>
> 1. Nodetool cfstats for both the tables.
> 2. Nodetool cfhistograms for both the tables.
> 3. Replication factor of the tables.
> 4. Consistency with which write requests are sent
> 5. Also the type of write queries for the table  if handy would also help
> (Light weight transactions or Batch writes or Prepared statements)
>
> Thanks
>
> On 2019/07/18 15:48:09, CPC <a...@gmail.com> wrote:
> > Hi all,>
> >
> > Our cassandra cluster consist of two dc and every dc we have 10 nodes.
> We>
> > are using DSE 5.1.12 (cassandra 3.11).We have a high local write latency
> on>
> > a single table. All other tables in our keyspace  have normal latencies>
> > like 0.02 msec,even tables that have more write tps and more data.
> Below>
> > you can find two table descriptions and their latencies.>
> > message_history_state have high local write latency. This is not node>
> > specific every node have this high local write latency for>
> > message_history_state. Have you ever see such a behavior or any clue
> why>
> > this could happen?>
> >
> >  CREATE TABLE tims."MESSAGE_HISTORY" (>
> > >     username text,>
> > >     date_partition text,>
> > >     jid text,>
> > >     sent_time timestamp,>
> > >     message_id text,>
> > >     stanza text,>
> > >     PRIMARY KEY ((username, date_partition), jid, sent_time,
> message_id)>
> > > ) WITH CLUSTERING ORDER BY (jid ASC, sent_time DESC, message_id ASC)>
> > >     AND bloom_filter_fp_chance = 0.01>
> > >     AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}>
> > >     AND comment = ''>
> > >     AND compaction = {'bucket_high': '1.5', 'bucket_low': '0.5',
> 'class':>
> > > 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',>
> > > 'enabled': 'true', 'max_threshold': '32', 'min_sstable_size': '50',>
> > > 'min_threshold': '4', 'tombstone_compaction_interval': '86400',>
> > > 'tombstone_threshold': '0.2', 'unchecked_tombstone_compaction':
> 'false'}>
> > >     AND compression = {'chunk_length_in_kb': '64', 'class':>
> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}>
> > >     AND crc_check_chance = 1.0>
> > >     AND dclocal_read_repair_chance = 0.0>
> > >     AND default_time_to_live = 0>
> > >     AND gc_grace_seconds = 86400>
> > >     AND max_index_interval = 2048>
> > >     AND memtable_flush_period_in_ms = 0>
> > >     AND min_index_interval = 128>
> > >     AND read_repair_chance = 0.0>
> > >     AND speculative_retry = '99PERCENTILE';>
> > >>
> > > CREATE TABLE tims."MESSAGE_HISTORY_STATE" (>
> > >     username text,>
> > >     date_partition text,>
> > >     message_id text,>
> > >     jid text,>
> > >     state text,>
> > >     sent_time timestamp,>
> > >     PRIMARY KEY ((username, date_partition), message_id, jid, state)>
> > > ) WITH CLUSTERING ORDER BY (message_id ASC, jid ASC, state ASC)>
> > >     AND bloom_filter_fp_chance = 0.01>
> > >     AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}>
> > >     AND comment = ''>
> > >     AND compaction = {'class':>
> > > 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',>
> > > 'max_threshold': '32', 'min_threshold': '4'}>
> > >     AND compression = {'chunk_length_in_kb': '64', 'class':>
> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}>
> > >     AND crc_check_chance = 1.0>
> > >     AND dclocal_read_repair_chance = 0.1>
> > >     AND default_time_to_live = 0>
> > >     AND gc_grace_seconds = 864000>
> > >     AND max_index_interval = 2048>
> > >     AND memtable_flush_period_in_ms = 0>
> > >     AND min_index_interval = 128>
> > >     AND read_repair_chance = 0.0>
> > >     AND speculative_retry = '99PERCENTILE';>
> > >>
> >
> > message_history Local write latency: 0.021 ms>
> > message_history_state Local write latency: 11.353 ms>
> >
> > Thanks in advance.>
> >
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> For additional commands, e-mail: user-h...@cassandra.apache.org
>
>

Reply via email to