On 23/2/26 21:25, Andrei Lepikhov wrote:
On 23/2/26 18:03, Attila Soki wrote:
So, let me discover a little more, but your PG14 explain could add more
details here.
It seems much more interesting than just a trivial accumulation of cost
estimation errors. Look:
...
-> Hash (cost=86.59..86.59 rows=8 width=67)
(actual time=0.136..0.136 rows=44.56 loops=21798)
Buckets: 2048 (originally 1024) Batches: 1 (originally 1) ...
-> Nested Loop (cost=1.12..86.59 rows=8 width=67)
(actual time=0.017..0.126 rows=44.56 loops=21798)
...
This hash table has been rescanned multiple times. And on each rescan,
it was rebuilt as well (the number of loops in the underlying Join was
also 21798). It is the first time I have seen such a query plan. And
discovering how rescan reckons in the cost model, this Hash table
rebuilding == subtree rescanning, you may find the following:
cost_rescan():
case T_HashJoin:
/*
* If it's a single-batch join, we don't need to rebuild the hash
* table during a rescan.
*/
if (((HashPath *) path)->num_batches == 1)
{
/* Startup cost is exactly the cost of hash table building */
*rescan_startup_cost = 0;
*rescan_total_cost = path->total_cost - path->startup_cost;
}
...
That means (if I read the code correctly) we don't take into account the
cost=86.59 of subtree rescanning and htab rebuilding at all!
So, it looks like a rare cost model bug.
To learn more, I still need your PG14 EXPLAIN. Can you also share your
SQL so we can understand which combination of SQL structures led to this
unusual query plan?
--
regards, Andrei Lepikhov,
pgEdge