Alvaro Herrera <alvhe...@2ndquadrant.com> writes:
> Tom Lane wrote:
>> Presumably what is happening is that the planner is switching from hash
>> to sort aggregation.

> I can't imagine that the server is avoiding hash aggregation on a 1MB
> work_mem limit for data that's a few dozen of bytes.  Is it really doing
> that?

Yup:

regression=# explain SELECT v,h, string_agg(i::text, E'\n') AS i FROM ctv_data
GROUP BY v, h ORDER BY h,v;
                               QUERY PLAN                               
------------------------------------------------------------------------
 Sort  (cost=33.87..34.37 rows=200 width=96)
   Sort Key: h, v
   ->  HashAggregate  (cost=23.73..26.23 rows=200 width=96)
         Group Key: h, v
         ->  Seq Scan on ctv_data  (cost=0.00..16.10 rows=610 width=68)
(5 rows)

regression=# set work_mem = '1MB';
SET
regression=# explain SELECT v,h, string_agg(i::text, E'\n') AS i FROM ctv_data
GROUP BY v, h ORDER BY h,v;
                               QUERY PLAN                               
------------------------------------------------------------------------
 GroupAggregate  (cost=44.32..55.97 rows=200 width=96)
   Group Key: h, v
   ->  Sort  (cost=44.32..45.85 rows=610 width=68)
         Sort Key: h, v
         ->  Seq Scan on ctv_data  (cost=0.00..16.10 rows=610 width=68)
(5 rows)

Now that you mention it, this does seem a bit odd, although I remember
that there's a pretty substantial fudge factor in there when we have
no statistics (which we don't in this example).  If I ANALYZE ctv_data
then it sticks to the hashagg plan all the way down to 64kB work_mem.

                        regards, tom lane


-- 
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

Reply via email to