On Thu, Aug 28, 2014 at 4:41 PM, Amit Kapila <amit.kapil...@gmail.com>
wrote:
>
> I have yet to collect data under varying loads, however I have
> collected performance data for 8GB shared buffers which shows
> reasonably good performance and scalability.
>
> I think the main part left for this patch is more data for various loads
> which I will share in next few days, however I think patch is ready for
> next round of review, so I will mark it as Needs Review.

I have collected more data with the patch.  I understand that you
have given more review comments due to which patch require
changes, however I think it will not effect the performance data
to a great extent and I have anyway taken the data, so sharing the
same.

> Performance Data:
> -------------------------------
>
> Configuration and Db Details
> IBM POWER-7 16 cores, 64 hardware threads
> RAM = 64GB
> Database Locale =C
> checkpoint_segments=256
> checkpoint_timeout    =15min
> scale factor = 3000
> Client Count = number of concurrent sessions and threads (ex. -c 8 -j 8)
> Duration of each individual run = 5mins
>
> All the data is in tps and taken using pgbench read-only load

Common configuration remains same as above.

Shared_Buffers = 500MB
  Client Count/Patch_Ver 8 16 32 64 128  HEAD 56248 100112 121341 81128
56552  Patch 59389 112483 157034 185740 166725

Shared_Buffers = 1GB
  Client Count/Patch_Ver 8 16 32 64 128  HEAD 56401 102557 121643 81686
57091  Patch 59361 114813 157528 188209 167752


Shared_Buffers = 14GB
  Client Count/Patch_Ver 8 16 32 64 128  HEAD 60059 110582 152051 130718
97014  Patch 61462 117377 169767 225803 229083


Shared_Buffers = 15GB
  Client Count/Patch_Ver 8 16 32 64 128  HEAD 60005 112928 153650 135203
36343  Patch 61345 115569 168767 226150 36985


Observations
---------------------
1. Performance improvement is upto 2~3 times for higher client
counts (64, 128).
2. For lower client count (8), we can see 2~5 % performance
improvement.
3. Overall, this improves the read scalability.
4. For lower number of shared buffers, we see that there is a minor
dip in tps even after patch (it might be that we can improve it by
tuning higher water mark for the number of buffers on freelist, I will
try this by varying high water mark).
5. For larger shared buffers (15GB), we can see that there is still a
dip at large client count, although situation is not bad as compare to
HEAD.  The reason is that at such high shared buffers and client count,
I/O starts happening because all the data can't be contained in RAM.

I will try to take some data for tpc-b load as well.  Kindly let me know
if you want to see data for some other configuration.


With Regards,
Amit Kapila.
EnterpriseDB: http://www.enterprisedb.com

Reply via email to