I have a single structured row as input with rate of 10K per seconds. Each row has 20 columns. Some queries should be answered on these inputs. Because most of queries needs different where, group by or orderby, The final data model ended up like this:
primary key for table of query1 : ((column1,column2),column3,column4) primary key for table of query2 : ((column3,column4),column2,column1) and so on I am aware of the limit in number of tables in cassandra data model (200 is warning and 500 would fail) Because for every input row i should do an insert in every table, the final write per seconds became big * big data!: write per seconds = 10K (input) * number of tables (queries) * replication factor The main question: am i in the right path? is this normal to have a table for every query even when the input rate is already so high? Shouldn't i use something like spark or hadoop upon instead of relying on bare datamodel Or event Hbase instead of cassandra? Sent using Zoho Mail