Re: Extract big data to file

2017-02-08 Thread Justin Cameron
Actually using BEGINTOKEN and ENDTOKEN will only give you what you want if you're using ByteOrderedPartitioner (not with the default murmur3). It also looks like *datetimestamp *is a clustering column so that suggestion probably wouldn't have applied anyway. On Wed, 8 Feb 2017 at 13:04 Justin

Re: Extract big data to file

2017-02-08 Thread Justin Cameron
Ideally you would have the program/Spark job that receives the data from Kafka write it to a text file as it writes each row to Cassandra - that way you don't need to query Cassandra at all. If you need to dump this data ad-hoc, rather than on a regular schedule, your best bet is to write some

Re: Extract big data to file

2017-02-08 Thread Kiril Menshikov
Did you try to receive data through the code? cqlsh probably not the right tool to fetch 360G. > On Feb 8, 2017, at 12:34, Cogumelos Maravilha > wrote: > > Hi list, > > My database stores data from Kafka. Using C* 3.0.10 > > In my cluster I'm using: > AND

Extract big data to file

2017-02-08 Thread Cogumelos Maravilha
Hi list, My database stores data from Kafka. Using C* 3.0.10 In my cluster I'm using: AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'} The result of extract one day of data uncompressed is around 360G. I've find these approaches: echo "SELECT kafka