Since I will only have 1 table with that many columns, and the other tables will be "normal" tables with max 30 columns, and the memory of 2K columns won't be that big, I'm gonna guess I'll be fine.
The data model is too dynamic, the alternative would be to create a table for each user which will have even more overhead since the number of users is in the several thousands/millions. On Thu, Sep 15, 2016 at 3:04 PM, DuyHai Doan <doanduy...@gmail.com> wrote: > There is no real limit in term of number of columns in a table, I would > say that the impact of having a lot of columns is the amount of meta data > C* needs to keep in memory for encoding/decoding each row. > > Now, if you have a table with 1000+ columns, the problem is probably your > data model... > > On Thu, Sep 15, 2016 at 2:59 PM, Dorian Hoxha <dorian.ho...@gmail.com> > wrote: > >> Is there alot of overhead with having a big number of columns in a table >> ? Not unbounded, but say, would 2000 be a problem(I think that's the >> maximum I'll need) ? >> >> Thank You >> > >