On Fri, Sep 23, 2016 at 3:12 AM, Dev Nop <devn...@gmail.com> wrote: > I’m storing thousands of independent documents each containing around 20k > rows. The larger the document, the more likely it is to be active with > inserts and updates (1000s/day). The most common read query is to get all > the rows for a single document (100s/day). >
How can the query be an order of magnitude less than the writes? Wouldn't anything doing an insert or update want to see the results of other people's inserts/updates about as frequently as they happen? > It will be supporting real-time collaboration but with strong-consistency > for a simple schema so not well-suited to dedicated "document databases" > that assume schema-less & eventual consistency. I won’t have great > hardware/budget so need to squeeze the most out of the least. > > My question is whether to put all documents into a single huge table or > partition by document? > > The documents are independent so its purely a performance question. Its > too many tables for postgresql partitioning support but I don’t get any > benefit from a master table and constraints. Handling partitioning in > application logic is effectively zero cost. > > I know that 1000s of tables is regarded as an anti-pattern but I can only > see the performance and maintenance benefits of one table per independent > document e.g. fast per-table vacuum, incremental schema updates, easy > future sharding. A monster table will require additional key columns and > indexes that don’t have any value beyond allowing the documents to sit in > the same table. > If you go the partitioned route, I would add the extra column anyway (but not an index on it), so that it is there if/when you need it. > > The only downsides seem to be the system level per-table overhead but I > only see that as a problem if I have a very long tail of tiny documents. > I'd rather solve that problem if it occurs than manage an > all-eggs-in-one-basket monster table. > > Is there anything significant I am missing in my reasoning? > If you use a reasonably modern version of PostgreSQL (say, >=9.4) , the overhead of having 1000s of tables should not be too large of a problem. When get into the 100,000 range, that it is likely to start being a problem. If you get to 1,000,000, you almost definitely have a problem. Cheers, Jeff