Another possible alternative is to use a single map column. On Thu, Sep 15, 2016 at 7:19 AM Dorian Hoxha <dorian.ho...@gmail.com> wrote:
> Since I will only have 1 table with that many columns, and the other > tables will be "normal" tables with max 30 columns, and the memory of 2K > columns won't be that big, I'm gonna guess I'll be fine. > > The data model is too dynamic, the alternative would be to create a table > for each user which will have even more overhead since the number of users > is in the several thousands/millions. > > > On Thu, Sep 15, 2016 at 3:04 PM, DuyHai Doan <doanduy...@gmail.com> wrote: > >> There is no real limit in term of number of columns in a table, I would >> say that the impact of having a lot of columns is the amount of meta data >> C* needs to keep in memory for encoding/decoding each row. >> >> Now, if you have a table with 1000+ columns, the problem is probably your >> data model... >> >> On Thu, Sep 15, 2016 at 2:59 PM, Dorian Hoxha <dorian.ho...@gmail.com> >> wrote: >> >>> Is there alot of overhead with having a big number of columns in a table >>> ? Not unbounded, but say, would 2000 be a problem(I think that's the >>> maximum I'll need) ? >>> >>> Thank You >>> >> >> >