Hello Kevin
Can you be more specific on the issue you're facing ? What is the table
design ? What kind of query are you doing ?
Regards
On Fri, May 30, 2014 at 7:10 AM, Kevin Burton bur...@spinn3r.com wrote:
I'm trying to grok this but I can't figure it out in CQL world.
I'd like to
On Thu, May 29, 2014 at 10:10 PM, Kevin Burton bur...@spinn3r.com wrote:
I'd like to efficiently page through a table via primary key.
This way I only involve one node at a time and the reads on disk are
This is only true if you use an Ordered Partitioner, which almost no one
does?
I
I think what you want is a clustering column”. When you model your data, you
specify “partition columns” which are synonymous with the old thrift style
“keys” and clustering columns. When creating your PRIMARY KEY, you specify the
partition column first then each subsequent column in the
The specific issue is I have a fairly large table, which is immutable, and
I need to get it in a form where it can be downloaded, page by page, via an
API.
This would involve reading the whole table.
I'd like to page through it by key order to efficiently read the rows to
minimize random reads.
Then the data model you chose is incorrect. As Rob Coli mentioned, you can not
page through partitions that are ordered unless you are using an ordered
partitioner. Your only option is to store the data differently. When using
Cassandra you have to remember to “model your queries, not your
Hello Kevin
One possible data model:
CREATE TABLE myLog(
day int //day format as MMdd,
date timeuuid,
log_message text,
PRIMARY_KEY(day,date)
);
For each day, you can query paging by date (timeuuid format). SELECT
log_message FROM myLog where day = 20140530 AND date... LIMIT xxx;
I'm trying to grok this but I can't figure it out in CQL world.
I'd like to efficiently page through a table via primary key.
This way I only involve one node at a time and the reads on disk are
contiguous.
I would have assumed it was a combination of pk and order by but that
doesn't seem to