A simpler approach might be to insert expiring columns into a 2nd CF with a TTL of one hour.
On Mon, Oct 4, 2010 at 5:12 AM, Utku Can Topçu <u...@topcu.gen.tr> wrote: > Hey All, > > I'm planning to run Map/Reduce on one of the ColumnFamilies. The keys are > formed in such a fashion that, they are indexed in descending order by time. > So I'll be analyzing the data for every hour iteratively. > > Since the current Hadoop integration does not support partial columnfamily > analysis. I feel that, I'll need to dump the data of the last hour and put > it to the hadoop cluster and do my analysis on the flat text file. > Do you think of any other "better" way of getting the data of a keyrange > into a hadoop cluster for analysis? > > Regards, > > Utku > > > -- Jonathan Ellis Project Chair, Apache Cassandra co-founder of Riptano, the source for professional Cassandra support http://riptano.com