Re: Compacting more than the actual used space

2018-11-05 Thread Pedro Gordo
Hi Alexander

Thanks. Using the compression ratio, the sizes check out.

Regarding the new values for compaction throughput, that explains it then.
We are using 2.1. :-)

Cheers
Pedro Gordo


On Mon, 5 Nov 2018 at 19:53, Alexander Dejanovski 
wrote:

> You can check cfstats to see what's the compression ratio.
> It's totally possible to have the values you're reporting as a compression
> ratio of 0.2 is quite common depending on the data you're storing
> (compressed size is then 20% of the original data).
>
> Compaction throughput changes are taken into account for running
> compactions starting with Cassandra 2.2 if I'm correct. Your compaction
> could be bound by cpu, not I/O in that case.
>
> Cheers
>
> Le lun. 5 nov. 2018 à 20:41, Pedro Gordo  a
> écrit :
>
>> Hi
>>
>> We have an ongoing compaction for roughly 2.5 TB, but "nodetool status"
>> reports a load of 1.09 TB. Even if we take into account that the load
>> presented by "nodetool status" is the compressed size, I very much doubt
>> that compression would work to reduce from 2.5 TB to 1.09.
>> We can also take into account that, even if this is the biggest table,
>> there are other tables in the system, so the 1.09 TB reported is not just
>> for the table being compacted.
>>
>> What could lead to results like this? We have 4 attached volumes for data
>> directories. Could this be a likely cause for such discrepancy?
>>
>> Bonus question: changing the compaction throughput to 0 (removing the
>> throttling), had no impacts in the current compaction. Do new compaction
>> throughput values only come into effect when a new compaction kicks in?
>>
>> Cheers
>>
>> Pedro Gordo
>>
> --
> -
> Alexander Dejanovski
> France
> @alexanderdeja
>
> Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
>


Compacting more than the actual used space

2018-11-05 Thread Pedro Gordo
Hi

We have an ongoing compaction for roughly 2.5 TB, but "nodetool status"
reports a load of 1.09 TB. Even if we take into account that the load
presented by "nodetool status" is the compressed size, I very much doubt
that compression would work to reduce from 2.5 TB to 1.09.
We can also take into account that, even if this is the biggest table,
there are other tables in the system, so the 1.09 TB reported is not just
for the table being compacted.

What could lead to results like this? We have 4 attached volumes for data
directories. Could this be a likely cause for such discrepancy?

Bonus question: changing the compaction throughput to 0 (removing the
throttling), had no impacts in the current compaction. Do new compaction
throughput values only come into effect when a new compaction kicks in?

Cheers
Pedro Gordo


Re: disable compaction if all data are read-only?

2016-04-08 Thread Pedro Gordo
Hi Yatong

My understanding is that if you have a table whichi read-only and hence
doesn't receive any writes, then no SSTables will be created, and hence, no
compaction will happen. What compaction strategy do you have on your table?

Best regards

Pedro Gordo

On 8 April 2016 at 10:42, Yatong Zhang <bluefl...@gmail.com> wrote:

> Hi there,
> I am wondering if it is possible to disable compaction when all my data
> are read-only?
>