Hi guys,

I am using rowkeys with a pattern like [minute]_[timestamp] because my
main use case is to read time ranges over a couple of hours and I want
to read in parallel from as many nodes in the cluster as possible,
thus, distributing the data in minute buckets across the cluster.

Problem now is that I am not sure how to do sequential reads (for
example all records between 11:10 and 12:00) and for defining such
time frames as input to my MapReduce jobs.

Any ideas?

Thanks,
Felix

Reply via email to