Actually using BEGINTOKEN and ENDTOKEN will only give you what you want if
you're using ByteOrderedPartitioner (not with the default murmur3). It also
looks like *datetimestamp *is a clustering column so that suggestion
probably wouldn't have applied anyway.
On Wed, 8 Feb 2017 at 13:04 Justin
Ideally you would have the program/Spark job that receives the data from
Kafka write it to a text file as it writes each row to Cassandra - that way
you don't need to query Cassandra at all.
If you need to dump this data ad-hoc, rather than on a regular schedule,
your best bet is to write some
Did you try to receive data through the code? cqlsh probably not the right tool
to fetch 360G.
> On Feb 8, 2017, at 12:34, Cogumelos Maravilha
> wrote:
>
> Hi list,
>
> My database stores data from Kafka. Using C* 3.0.10
>
> In my cluster I'm using:
> AND
Hi list,
My database stores data from Kafka. Using C* 3.0.10
In my cluster I'm using:
AND compression = {'sstable_compression':
'org.apache.cassandra.io.compress.LZ4Compressor'}
The result of extract one day of data uncompressed is around 360G.
I've find these approaches:
echo "SELECT kafka