Thanks a lot to share the result.

Boa Sorte.
;-)
Take care.
Petris Silva

On Fri, Jul 21, 2017 at 12:19 PM, Felipe Esteves <
felipe.este...@b2wdigital.com> wrote:

> Hi, Petrus,
>
> Seems we've solved the problem, but it wasn't relationed to repair the
> cluster or disk latency.
> I've increased the memory available for Cassandra from 16GB to 24GB and
> the performance was much improved!
> The main symptom we've observed in Opscenter was a significantly decrease
> in total compactions graph.
>
> Felipe Esteves
>
> Tecnologia
>
> felipe.este...@b2wdigital.com <seu.em...@b2wdigital.com>
>
>
>
> 2017-07-15 3:23 GMT-03:00 Petrus Gomes <petru...@gmail.com>:
>
>> Hi Felipe,
>>
>> Yes, try it and let us know how it goes.
>>
>> Thanks,
>> Petrus Silva.
>>
>> On Fri, Jul 14, 2017 at 11:37 AM, Felipe Esteves <
>> felipe.este...@b2wdigital.com> wrote:
>>
>>> Hi Petrus, thanks for the feedback.
>>>
>>> I couldn't found the percent repaired in nodetool info, C* version is
>>> 2.1.8, maybe it's something newer than that?
>>>
>>> I'm analyzing this thread about num_token.
>>>
>>> Compaction is "compaction_throughput_mb_per_sec: 16", I don't get
>>> pending compactions in Opscenter.
>>>
>>> One point I've noticed, is that Opscenter show "OS: Disk Latency" max
>>> with high values when the problem occurs, but it doesn't reflect in server
>>> directly monitoring, in these tools the IO and latency of disks seems ok.
>>> But seems to me that "read repair attempted" is a bit high, maybe it
>>> will explain the latency in reads. I will try to run a repair on cluster to
>>> see how it goes.
>>>
>>> Felipe Esteves
>>>
>>> Tecnologia
>>>
>>> felipe.este...@b2wdigital.com <seu.em...@b2wdigital.com>
>>>
>>> Tel.: (21) 3504-7162 ramal 57162
>>>
>>> Skype: felipe2esteves
>>>
>>> 2017-07-13 15:02 GMT-03:00 Petrus Gomes <petru...@gmail.com>:
>>>
>>>> How is your Percent Repaired  when you run " nodetool info" ?
>>>>
>>>> Search for :
>>>> "reduced num_token = improved performance ??" topic.
>>>> The people were discussing that.
>>>>
>>>> How is your compaction is configured?
>>>>
>>>> Could you run the same process in command line to have a measurement?
>>>>
>>>> Thanks,
>>>> Petrus Silva
>>>>
>>>>
>>>>
>>>> On Thu, Jul 13, 2017 at 7:49 AM, Felipe Esteves <
>>>> felipe.este...@b2wdigital.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I have a Cassandra 2.1 cluster running on AWS that receives high read
>>>>> loads, jumping from 100k requests to 400k requests, for example. Then it
>>>>> normalizes and later cames another high throughput.
>>>>>
>>>>> To the application, it appears that Cassandra is slow. However, cpu
>>>>> and disk use is ok in every instance, row cache is enabled and with almost
>>>>> 100% hit rate.
>>>>>
>>>>> The logs from Cassandra instances doesn't have any errors, nor
>>>>> tombstone messages or something liked that. It's mostly compactions and
>>>>> G1GC operations.
>>>>>
>>>>> Any hints on where to investigate more?
>>>>>
>>>>>
>>>>> Felipe Esteves
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>>> ------------------------------
>>>>
>>>> Esta mensagem pode conter informações confidenciais e somente o
>>>> indivíduo ou entidade a quem foi destinada pode utilizá-la. A transmissão
>>>> incorreta da mensagem não acarreta a perda de sua confidencialidade. Caso
>>>> esta mensagem tenha sido recebida por engano, solicitamos que o fato seja
>>>> comunicado ao remetente e que a mensagem seja eliminada de seu sistema
>>>> imediatamente. É vedado a qualquer pessoa que não seja o destinatário usar,
>>>> revelar, distribuir ou copiar qualquer parte desta mensagem. Ambiente de
>>>> comunicação sujeito a monitoramento.
>>>>
>>>> This message may include confidential information and only the intended
>>>> addresses have the right to use it as is, or any part of it. A wrong
>>>> transmission does not break its confidentiality. If you've received it
>>>> because of a mistake or erroneous transmission, please notify the sender
>>>> and delete it from your system immediately. This communication environment
>>>> is controlled and monitored.
>>>>
>>>> B2W Digital
>>>>
>>>>
>>>>
>>>
>>>
>>>
>>>
>>
>
>
>

Reply via email to