Re: Cassandra seems slow when having many read operations

2017-07-24 Thread Jeff Jirsa


On 2017-07-13 07:49 (-0700), Felipe Esteves  
wrote: 
> Hi,
> 
> I have a Cassandra 2.1 cluster running on AWS that receives high read
> loads, jumping from 100k requests to 400k requests, for example. Then it
> normalizes and later cames another high throughput.
> 
> To the application, it appears that Cassandra is slow. However, cpu and
> disk use is ok in every instance, row cache is enabled and with almost 100%
> hit rate.
> 
> The logs from Cassandra instances doesn't have any errors, nor tombstone
> messages or something liked that. It's mostly compactions and G1GC
> operations.
> 
> Any hints on where to investigate more?
> 

What compaction strategy are you using?
How many writes/second are you doing?
Is compaction keeping up? 

STCS has a little-known feature called "row lifting", where it can force 
re-compaction of data to avoid merging results from many sstables - this can 
occasionally account for a lot of extra disk IO (and gc) in high read clusters.

Apart from that, how many sstables are you touching per read, or is it all 
coming straight out of row cache?

If you're really seeing 90% row hit rates (good!), you may want to make sure 
you're caching enough of the partition that you don't have to read past what's 
cached - make sure your rows per partition setting is high enough (think it 
defaults to 100).




-
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org



Re: Cassandra seems slow when having many read operations

2017-07-23 Thread Nate McCall
On Sat, Jul 15, 2017 at 6:37 AM, Felipe Esteves <
felipe.este...@b2wdigital.com> wrote:

>
> One point I've noticed, is that Opscenter show "OS: Disk Latency" max with
> high values when the problem occurs, but it doesn't reflect in server
> directly monitoring, in these tools the IO and latency of disks seems ok.
>

YMMV, but I've seen something like this due to an issue balancing IRQ
interrupts on older 3 series kernels. Check the output of 'cat
/proc/interrupts' and make sure the interrupts for the disks and network
driver(s) particularly are not contending. This article explains the issue
in detail (as well as how to fix it):
http://www.alexonlinux.com/smp-affinity-and-proper-interrupt-handling-in-linux


Re: Cassandra seems slow when having many read operations

2017-07-22 Thread benjamin roth
Chunk size:
For us it made a 20x difference in read io. But it depends a lot on the use
case.

Am 22.07.2017 08:32 schrieb "Fay Hou [Storage Service] ­" <
fay...@coupang.com>:

> Hey Felipe:
>
> When you say increased memory from 16GB to 24GB, I think you meant you
> increased heap to 24GB. do you use cms or g1gc?
>  did you change any other parameters?
> As for the chunk size, we found change 64kb to 16kb didn't make a
> difference in low key cache rate environment
>
>
>
> On Fri, Jul 21, 2017 at 9:27 PM, benjamin roth  wrote:
>
>> Apart from all that you can try to reduce the compression chunk size from
>> the default 64kb to 16kb or even down to 4kb. This can help a lot if your
>> read io on disk is very high and the page cache is not efficient.
>>
>> Am 21.07.2017 23:03 schrieb "Petrus Gomes" :
>>
>>> Thanks a lot to share the result.
>>>
>>> Boa Sorte.
>>> ;-)
>>> Take care.
>>> Petris Silva
>>>
>>> On Fri, Jul 21, 2017 at 12:19 PM, Felipe Esteves <
>>> felipe.este...@b2wdigital.com> wrote:
>>>
 Hi, Petrus,

 Seems we've solved the problem, but it wasn't relationed to repair the
 cluster or disk latency.
 I've increased the memory available for Cassandra from 16GB to 24GB and
 the performance was much improved!
 The main symptom we've observed in Opscenter was a
 significantly decrease in total compactions graph.

 Felipe Esteves

 Tecnologia

 felipe.este...@b2wdigital.com 



 2017-07-15 3:23 GMT-03:00 Petrus Gomes :

> Hi Felipe,
>
> Yes, try it and let us know how it goes.
>
> Thanks,
> Petrus Silva.
>
> On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
> felipe.este...@b2wdigital.com> wrote:
>
>> Hi Petrus, thanks for the feedback.
>>
>> I couldn't found the percent repaired in nodetool info, C* version is
>> 2.1.8, maybe it's something newer than that?
>>
>> I'm analyzing this thread about num_token.
>>
>> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
>> pending compactions in Opscenter.
>>
>> One point I've noticed, is that Opscenter show "OS: Disk Latency" max
>> with high values when the problem occurs, but it doesn't reflect in 
>> server
>> directly monitoring, in these tools the IO and latency of disks seems ok.
>> But seems to me that "read repair attempted" is a bit high, maybe it
>> will explain the latency in reads. I will try to run a repair on cluster 
>> to
>> see how it goes.
>>
>> Felipe Esteves
>>
>> Tecnologia
>>
>> felipe.este...@b2wdigital.com 
>>
>> Tel.: (21) 3504-7162 ramal 57162
>>
>> Skype: felipe2esteves
>>
>> 2017-07-13 15:02 GMT-03:00 Petrus Gomes :
>>
>>> How is your Percent Repaired  when you run " nodetool info" ?
>>>
>>> Search for :
>>> "reduced num_token = improved performance ??" topic.
>>> The people were discussing that.
>>>
>>> How is your compaction is configured?
>>>
>>> Could you run the same process in command line to have a measurement?
>>>
>>> Thanks,
>>> Petrus Silva
>>>
>>>
>>>
>>> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
>>> felipe.este...@b2wdigital.com> wrote:
>>>
 Hi,

 I have a Cassandra 2.1 cluster running on AWS that receives high
 read loads, jumping from 100k requests to 400k requests, for example. 
 Then
 it normalizes and later cames another high throughput.

 To the application, it appears that Cassandra is slow. However, cpu
 and disk use is ok in every instance, row cache is enabled and with 
 almost
 100% hit rate.

 The logs from Cassandra instances doesn't have any errors, nor
 tombstone messages or something liked that. It's mostly compactions and
 G1GC operations.

 Any hints on where to investigate more?


 Felipe Esteves





>>>
>>> --
>>>
>>> Esta mensagem pode conter informações confidenciais e somente o
>>> indivíduo ou entidade a quem foi destinada pode utilizá-la. A 
>>> transmissão
>>> incorreta da mensagem não acarreta a perda de sua confidencialidade. 
>>> Caso
>>> esta mensagem tenha sido recebida por engano, solicitamos que o fato 
>>> seja
>>> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
>>> imediatamente. É vedado a qualquer pessoa que não seja o destinatário 
>>> usar,
>>> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
>>> comunicação sujeito a monitoramento.
>>>
>>> This message may include confidential 

Re: Cassandra seems slow when having many read operations

2017-07-22 Thread Fay Hou [Storage Service] ­
Hey Felipe:

When you say increased memory from 16GB to 24GB, I think you meant you
increased heap to 24GB. do you use cms or g1gc?
 did you change any other parameters?
As for the chunk size, we found change 64kb to 16kb didn't make a
difference in low key cache rate environment



On Fri, Jul 21, 2017 at 9:27 PM, benjamin roth  wrote:

> Apart from all that you can try to reduce the compression chunk size from
> the default 64kb to 16kb or even down to 4kb. This can help a lot if your
> read io on disk is very high and the page cache is not efficient.
>
> Am 21.07.2017 23:03 schrieb "Petrus Gomes" :
>
>> Thanks a lot to share the result.
>>
>> Boa Sorte.
>> ;-)
>> Take care.
>> Petris Silva
>>
>> On Fri, Jul 21, 2017 at 12:19 PM, Felipe Esteves <
>> felipe.este...@b2wdigital.com> wrote:
>>
>>> Hi, Petrus,
>>>
>>> Seems we've solved the problem, but it wasn't relationed to repair the
>>> cluster or disk latency.
>>> I've increased the memory available for Cassandra from 16GB to 24GB and
>>> the performance was much improved!
>>> The main symptom we've observed in Opscenter was a
>>> significantly decrease in total compactions graph.
>>>
>>> Felipe Esteves
>>>
>>> Tecnologia
>>>
>>> felipe.este...@b2wdigital.com 
>>>
>>>
>>>
>>> 2017-07-15 3:23 GMT-03:00 Petrus Gomes :
>>>
 Hi Felipe,

 Yes, try it and let us know how it goes.

 Thanks,
 Petrus Silva.

 On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
 felipe.este...@b2wdigital.com> wrote:

> Hi Petrus, thanks for the feedback.
>
> I couldn't found the percent repaired in nodetool info, C* version is
> 2.1.8, maybe it's something newer than that?
>
> I'm analyzing this thread about num_token.
>
> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
> pending compactions in Opscenter.
>
> One point I've noticed, is that Opscenter show "OS: Disk Latency" max
> with high values when the problem occurs, but it doesn't reflect in server
> directly monitoring, in these tools the IO and latency of disks seems ok.
> But seems to me that "read repair attempted" is a bit high, maybe it
> will explain the latency in reads. I will try to run a repair on cluster 
> to
> see how it goes.
>
> Felipe Esteves
>
> Tecnologia
>
> felipe.este...@b2wdigital.com 
>
> Tel.: (21) 3504-7162 ramal 57162
>
> Skype: felipe2esteves
>
> 2017-07-13 15:02 GMT-03:00 Petrus Gomes :
>
>> How is your Percent Repaired  when you run " nodetool info" ?
>>
>> Search for :
>> "reduced num_token = improved performance ??" topic.
>> The people were discussing that.
>>
>> How is your compaction is configured?
>>
>> Could you run the same process in command line to have a measurement?
>>
>> Thanks,
>> Petrus Silva
>>
>>
>>
>> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
>> felipe.este...@b2wdigital.com> wrote:
>>
>>> Hi,
>>>
>>> I have a Cassandra 2.1 cluster running on AWS that receives high
>>> read loads, jumping from 100k requests to 400k requests, for example. 
>>> Then
>>> it normalizes and later cames another high throughput.
>>>
>>> To the application, it appears that Cassandra is slow. However, cpu
>>> and disk use is ok in every instance, row cache is enabled and with 
>>> almost
>>> 100% hit rate.
>>>
>>> The logs from Cassandra instances doesn't have any errors, nor
>>> tombstone messages or something liked that. It's mostly compactions and
>>> G1GC operations.
>>>
>>> Any hints on where to investigate more?
>>>
>>>
>>> Felipe Esteves
>>>
>>>
>>>
>>>
>>>
>>
>> --
>>
>> Esta mensagem pode conter informações confidenciais e somente o
>> indivíduo ou entidade a quem foi destinada pode utilizá-la. A transmissão
>> incorreta da mensagem não acarreta a perda de sua confidencialidade. Caso
>> esta mensagem tenha sido recebida por engano, solicitamos que o fato seja
>> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
>> imediatamente. É vedado a qualquer pessoa que não seja o destinatário 
>> usar,
>> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
>> comunicação sujeito a monitoramento.
>>
>> This message may include confidential information and only the
>> intended addresses have the right to use it as is, or any part of it. A
>> wrong transmission does not break its confidentiality. If you've received
>> it because of a mistake or erroneous transmission, please notify the 
>> sender
>> and delete it from your system immediately. This communication 
>> environment

Re: Cassandra seems slow when having many read operations

2017-07-21 Thread benjamin roth
Apart from all that you can try to reduce the compression chunk size from
the default 64kb to 16kb or even down to 4kb. This can help a lot if your
read io on disk is very high and the page cache is not efficient.

Am 21.07.2017 23:03 schrieb "Petrus Gomes" :

> Thanks a lot to share the result.
>
> Boa Sorte.
> ;-)
> Take care.
> Petris Silva
>
> On Fri, Jul 21, 2017 at 12:19 PM, Felipe Esteves <
> felipe.este...@b2wdigital.com> wrote:
>
>> Hi, Petrus,
>>
>> Seems we've solved the problem, but it wasn't relationed to repair the
>> cluster or disk latency.
>> I've increased the memory available for Cassandra from 16GB to 24GB and
>> the performance was much improved!
>> The main symptom we've observed in Opscenter was a significantly decrease
>> in total compactions graph.
>>
>> Felipe Esteves
>>
>> Tecnologia
>>
>> felipe.este...@b2wdigital.com 
>>
>>
>>
>> 2017-07-15 3:23 GMT-03:00 Petrus Gomes :
>>
>>> Hi Felipe,
>>>
>>> Yes, try it and let us know how it goes.
>>>
>>> Thanks,
>>> Petrus Silva.
>>>
>>> On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
>>> felipe.este...@b2wdigital.com> wrote:
>>>
 Hi Petrus, thanks for the feedback.

 I couldn't found the percent repaired in nodetool info, C* version is
 2.1.8, maybe it's something newer than that?

 I'm analyzing this thread about num_token.

 Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
 pending compactions in Opscenter.

 One point I've noticed, is that Opscenter show "OS: Disk Latency" max
 with high values when the problem occurs, but it doesn't reflect in server
 directly monitoring, in these tools the IO and latency of disks seems ok.
 But seems to me that "read repair attempted" is a bit high, maybe it
 will explain the latency in reads. I will try to run a repair on cluster to
 see how it goes.

 Felipe Esteves

 Tecnologia

 felipe.este...@b2wdigital.com 

 Tel.: (21) 3504-7162 ramal 57162

 Skype: felipe2esteves

 2017-07-13 15:02 GMT-03:00 Petrus Gomes :

> How is your Percent Repaired  when you run " nodetool info" ?
>
> Search for :
> "reduced num_token = improved performance ??" topic.
> The people were discussing that.
>
> How is your compaction is configured?
>
> Could you run the same process in command line to have a measurement?
>
> Thanks,
> Petrus Silva
>
>
>
> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
> felipe.este...@b2wdigital.com> wrote:
>
>> Hi,
>>
>> I have a Cassandra 2.1 cluster running on AWS that receives high read
>> loads, jumping from 100k requests to 400k requests, for example. Then it
>> normalizes and later cames another high throughput.
>>
>> To the application, it appears that Cassandra is slow. However, cpu
>> and disk use is ok in every instance, row cache is enabled and with 
>> almost
>> 100% hit rate.
>>
>> The logs from Cassandra instances doesn't have any errors, nor
>> tombstone messages or something liked that. It's mostly compactions and
>> G1GC operations.
>>
>> Any hints on where to investigate more?
>>
>>
>> Felipe Esteves
>>
>>
>>
>>
>>
>
> --
>
> Esta mensagem pode conter informações confidenciais e somente o
> indivíduo ou entidade a quem foi destinada pode utilizá-la. A transmissão
> incorreta da mensagem não acarreta a perda de sua confidencialidade. Caso
> esta mensagem tenha sido recebida por engano, solicitamos que o fato seja
> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
> imediatamente. É vedado a qualquer pessoa que não seja o destinatário 
> usar,
> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
> comunicação sujeito a monitoramento.
>
> This message may include confidential information and only the
> intended addresses have the right to use it as is, or any part of it. A
> wrong transmission does not break its confidentiality. If you've received
> it because of a mistake or erroneous transmission, please notify the 
> sender
> and delete it from your system immediately. This communication environment
> is controlled and monitored.
>
> B2W Digital
>
>
>




>>>
>>
>>
>>
>


Re: Cassandra seems slow when having many read operations

2017-07-21 Thread Petrus Gomes
Thanks a lot to share the result.

Boa Sorte.
;-)
Take care.
Petris Silva

On Fri, Jul 21, 2017 at 12:19 PM, Felipe Esteves <
felipe.este...@b2wdigital.com> wrote:

> Hi, Petrus,
>
> Seems we've solved the problem, but it wasn't relationed to repair the
> cluster or disk latency.
> I've increased the memory available for Cassandra from 16GB to 24GB and
> the performance was much improved!
> The main symptom we've observed in Opscenter was a significantly decrease
> in total compactions graph.
>
> Felipe Esteves
>
> Tecnologia
>
> felipe.este...@b2wdigital.com 
>
>
>
> 2017-07-15 3:23 GMT-03:00 Petrus Gomes :
>
>> Hi Felipe,
>>
>> Yes, try it and let us know how it goes.
>>
>> Thanks,
>> Petrus Silva.
>>
>> On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
>> felipe.este...@b2wdigital.com> wrote:
>>
>>> Hi Petrus, thanks for the feedback.
>>>
>>> I couldn't found the percent repaired in nodetool info, C* version is
>>> 2.1.8, maybe it's something newer than that?
>>>
>>> I'm analyzing this thread about num_token.
>>>
>>> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
>>> pending compactions in Opscenter.
>>>
>>> One point I've noticed, is that Opscenter show "OS: Disk Latency" max
>>> with high values when the problem occurs, but it doesn't reflect in server
>>> directly monitoring, in these tools the IO and latency of disks seems ok.
>>> But seems to me that "read repair attempted" is a bit high, maybe it
>>> will explain the latency in reads. I will try to run a repair on cluster to
>>> see how it goes.
>>>
>>> Felipe Esteves
>>>
>>> Tecnologia
>>>
>>> felipe.este...@b2wdigital.com 
>>>
>>> Tel.: (21) 3504-7162 ramal 57162
>>>
>>> Skype: felipe2esteves
>>>
>>> 2017-07-13 15:02 GMT-03:00 Petrus Gomes :
>>>
 How is your Percent Repaired  when you run " nodetool info" ?

 Search for :
 "reduced num_token = improved performance ??" topic.
 The people were discussing that.

 How is your compaction is configured?

 Could you run the same process in command line to have a measurement?

 Thanks,
 Petrus Silva



 On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
 felipe.este...@b2wdigital.com> wrote:

> Hi,
>
> I have a Cassandra 2.1 cluster running on AWS that receives high read
> loads, jumping from 100k requests to 400k requests, for example. Then it
> normalizes and later cames another high throughput.
>
> To the application, it appears that Cassandra is slow. However, cpu
> and disk use is ok in every instance, row cache is enabled and with almost
> 100% hit rate.
>
> The logs from Cassandra instances doesn't have any errors, nor
> tombstone messages or something liked that. It's mostly compactions and
> G1GC operations.
>
> Any hints on where to investigate more?
>
>
> Felipe Esteves
>
>
>
>
>

 --

 Esta mensagem pode conter informações confidenciais e somente o
 indivíduo ou entidade a quem foi destinada pode utilizá-la. A transmissão
 incorreta da mensagem não acarreta a perda de sua confidencialidade. Caso
 esta mensagem tenha sido recebida por engano, solicitamos que o fato seja
 comunicado ao remetente e que a mensagem seja eliminada de seu sistema
 imediatamente. É vedado a qualquer pessoa que não seja o destinatário usar,
 revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
 comunicação sujeito a monitoramento.

 This message may include confidential information and only the intended
 addresses have the right to use it as is, or any part of it. A wrong
 transmission does not break its confidentiality. If you've received it
 because of a mistake or erroneous transmission, please notify the sender
 and delete it from your system immediately. This communication environment
 is controlled and monitored.

 B2W Digital



>>>
>>>
>>>
>>>
>>
>
>
>


Re: Cassandra seems slow when having many read operations

2017-07-21 Thread Felipe Esteves
Hi, Petrus,

Seems we've solved the problem, but it wasn't relationed to repair the
cluster or disk latency.
I've increased the memory available for Cassandra from 16GB to 24GB and the
performance was much improved!
The main symptom we've observed in Opscenter was a significantly decrease
in total compactions graph.

Felipe Esteves

Tecnologia

felipe.este...@b2wdigital.com 



2017-07-15 3:23 GMT-03:00 Petrus Gomes :

> Hi Felipe,
>
> Yes, try it and let us know how it goes.
>
> Thanks,
> Petrus Silva.
>
> On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
> felipe.este...@b2wdigital.com> wrote:
>
>> Hi Petrus, thanks for the feedback.
>>
>> I couldn't found the percent repaired in nodetool info, C* version is
>> 2.1.8, maybe it's something newer than that?
>>
>> I'm analyzing this thread about num_token.
>>
>> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
>> pending compactions in Opscenter.
>>
>> One point I've noticed, is that Opscenter show "OS: Disk Latency" max
>> with high values when the problem occurs, but it doesn't reflect in server
>> directly monitoring, in these tools the IO and latency of disks seems ok.
>> But seems to me that "read repair attempted" is a bit high, maybe it will
>> explain the latency in reads. I will try to run a repair on cluster to see
>> how it goes.
>>
>> Felipe Esteves
>>
>> Tecnologia
>>
>> felipe.este...@b2wdigital.com 
>>
>> Tel.: (21) 3504-7162 ramal 57162
>>
>> Skype: felipe2esteves
>>
>> 2017-07-13 15:02 GMT-03:00 Petrus Gomes :
>>
>>> How is your Percent Repaired  when you run " nodetool info" ?
>>>
>>> Search for :
>>> "reduced num_token = improved performance ??" topic.
>>> The people were discussing that.
>>>
>>> How is your compaction is configured?
>>>
>>> Could you run the same process in command line to have a measurement?
>>>
>>> Thanks,
>>> Petrus Silva
>>>
>>>
>>>
>>> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
>>> felipe.este...@b2wdigital.com> wrote:
>>>
 Hi,

 I have a Cassandra 2.1 cluster running on AWS that receives high read
 loads, jumping from 100k requests to 400k requests, for example. Then it
 normalizes and later cames another high throughput.

 To the application, it appears that Cassandra is slow. However, cpu and
 disk use is ok in every instance, row cache is enabled and with almost 100%
 hit rate.

 The logs from Cassandra instances doesn't have any errors, nor
 tombstone messages or something liked that. It's mostly compactions and
 G1GC operations.

 Any hints on where to investigate more?


 Felipe Esteves





>>>
>>> --
>>>
>>> Esta mensagem pode conter informações confidenciais e somente o
>>> indivíduo ou entidade a quem foi destinada pode utilizá-la. A transmissão
>>> incorreta da mensagem não acarreta a perda de sua confidencialidade. Caso
>>> esta mensagem tenha sido recebida por engano, solicitamos que o fato seja
>>> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
>>> imediatamente. É vedado a qualquer pessoa que não seja o destinatário usar,
>>> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
>>> comunicação sujeito a monitoramento.
>>>
>>> This message may include confidential information and only the intended
>>> addresses have the right to use it as is, or any part of it. A wrong
>>> transmission does not break its confidentiality. If you've received it
>>> because of a mistake or erroneous transmission, please notify the sender
>>> and delete it from your system immediately. This communication environment
>>> is controlled and monitored.
>>>
>>> B2W Digital
>>>
>>>
>>>
>>
>>
>>
>>
>

-- 



Re: Cassandra seems slow when having many read operations

2017-07-15 Thread Petrus Gomes
Hi Felipe,

Yes, try it and let us know how it goes.

Thanks,
Petrus Silva.

On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
felipe.este...@b2wdigital.com> wrote:

> Hi Petrus, thanks for the feedback.
>
> I couldn't found the percent repaired in nodetool info, C* version is
> 2.1.8, maybe it's something newer than that?
>
> I'm analyzing this thread about num_token.
>
> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get pending
> compactions in Opscenter.
>
> One point I've noticed, is that Opscenter show "OS: Disk Latency" max with
> high values when the problem occurs, but it doesn't reflect in server
> directly monitoring, in these tools the IO and latency of disks seems ok.
> But seems to me that "read repair attempted" is a bit high, maybe it will
> explain the latency in reads. I will try to run a repair on cluster to see
> how it goes.
>
> Felipe Esteves
>
> Tecnologia
>
> felipe.este...@b2wdigital.com 
>
> Tel.: (21) 3504-7162 ramal 57162
>
> Skype: felipe2esteves
>
> 2017-07-13 15:02 GMT-03:00 Petrus Gomes :
>
>> How is your Percent Repaired  when you run " nodetool info" ?
>>
>> Search for :
>> "reduced num_token = improved performance ??" topic.
>> The people were discussing that.
>>
>> How is your compaction is configured?
>>
>> Could you run the same process in command line to have a measurement?
>>
>> Thanks,
>> Petrus Silva
>>
>>
>>
>> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
>> felipe.este...@b2wdigital.com> wrote:
>>
>>> Hi,
>>>
>>> I have a Cassandra 2.1 cluster running on AWS that receives high read
>>> loads, jumping from 100k requests to 400k requests, for example. Then it
>>> normalizes and later cames another high throughput.
>>>
>>> To the application, it appears that Cassandra is slow. However, cpu and
>>> disk use is ok in every instance, row cache is enabled and with almost 100%
>>> hit rate.
>>>
>>> The logs from Cassandra instances doesn't have any errors, nor tombstone
>>> messages or something liked that. It's mostly compactions and G1GC
>>> operations.
>>>
>>> Any hints on where to investigate more?
>>>
>>>
>>> Felipe Esteves
>>>
>>>
>>>
>>>
>>>
>>
>> --
>>
>> Esta mensagem pode conter informações confidenciais e somente o indivíduo
>> ou entidade a quem foi destinada pode utilizá-la. A transmissão incorreta
>> da mensagem não acarreta a perda de sua confidencialidade. Caso esta
>> mensagem tenha sido recebida por engano, solicitamos que o fato seja
>> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
>> imediatamente. É vedado a qualquer pessoa que não seja o destinatário usar,
>> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
>> comunicação sujeito a monitoramento.
>>
>> This message may include confidential information and only the intended
>> addresses have the right to use it as is, or any part of it. A wrong
>> transmission does not break its confidentiality. If you've received it
>> because of a mistake or erroneous transmission, please notify the sender
>> and delete it from your system immediately. This communication environment
>> is controlled and monitored.
>>
>> B2W Digital
>>
>>
>>
>
>
>
>


Re: Cassandra seems slow when having many read operations

2017-07-14 Thread Felipe Esteves
Hi Petrus, thanks for the feedback.

I couldn't found the percent repaired in nodetool info, C* version is
2.1.8, maybe it's something newer than that?

I'm analyzing this thread about num_token.

Compaction is "compaction_throughput_mb_per_sec: 16", I don't get pending
compactions in Opscenter.

One point I've noticed, is that Opscenter show "OS: Disk Latency" max with
high values when the problem occurs, but it doesn't reflect in server
directly monitoring, in these tools the IO and latency of disks seems ok.
But seems to me that "read repair attempted" is a bit high, maybe it will
explain the latency in reads. I will try to run a repair on cluster to see
how it goes.

Felipe Esteves

Tecnologia

felipe.este...@b2wdigital.com 

Tel.: (21) 3504-7162 ramal 57162

Skype: felipe2esteves

2017-07-13 15:02 GMT-03:00 Petrus Gomes :

> How is your Percent Repaired  when you run " nodetool info" ?
>
> Search for :
> "reduced num_token = improved performance ??" topic.
> The people were discussing that.
>
> How is your compaction is configured?
>
> Could you run the same process in command line to have a measurement?
>
> Thanks,
> Petrus Silva
>
>
>
> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
> felipe.este...@b2wdigital.com> wrote:
>
>> Hi,
>>
>> I have a Cassandra 2.1 cluster running on AWS that receives high read
>> loads, jumping from 100k requests to 400k requests, for example. Then it
>> normalizes and later cames another high throughput.
>>
>> To the application, it appears that Cassandra is slow. However, cpu and
>> disk use is ok in every instance, row cache is enabled and with almost 100%
>> hit rate.
>>
>> The logs from Cassandra instances doesn't have any errors, nor tombstone
>> messages or something liked that. It's mostly compactions and G1GC
>> operations.
>>
>> Any hints on where to investigate more?
>>
>>
>> Felipe Esteves
>>
>>
>>
>>
>>
>
> --
>
> Esta mensagem pode conter informações confidenciais e somente o indivíduo
> ou entidade a quem foi destinada pode utilizá-la. A transmissão incorreta
> da mensagem não acarreta a perda de sua confidencialidade. Caso esta
> mensagem tenha sido recebida por engano, solicitamos que o fato seja
> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
> imediatamente. É vedado a qualquer pessoa que não seja o destinatário usar,
> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
> comunicação sujeito a monitoramento.
>
> This message may include confidential information and only the intended
> addresses have the right to use it as is, or any part of it. A wrong
> transmission does not break its confidentiality. If you've received it
> because of a mistake or erroneous transmission, please notify the sender
> and delete it from your system immediately. This communication environment
> is controlled and monitored.
>
> B2W Digital
>
>
>

-- 



Re: Cassandra seems slow when having many read operations

2017-07-13 Thread Petrus Gomes
How is your Percent Repaired  when you run " nodetool info" ?

Search for :
"reduced num_token = improved performance ??" topic.
The people were discussing that.

How is your compaction is configured?

Could you run the same process in command line to have a measurement?

Thanks,
Petrus Silva



On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
felipe.este...@b2wdigital.com> wrote:

> Hi,
>
> I have a Cassandra 2.1 cluster running on AWS that receives high read
> loads, jumping from 100k requests to 400k requests, for example. Then it
> normalizes and later cames another high throughput.
>
> To the application, it appears that Cassandra is slow. However, cpu and
> disk use is ok in every instance, row cache is enabled and with almost 100%
> hit rate.
>
> The logs from Cassandra instances doesn't have any errors, nor tombstone
> messages or something liked that. It's mostly compactions and G1GC
> operations.
>
> Any hints on where to investigate more?
>
>
> Felipe Esteves
>
>
>
>
>