Re: SLOW query with aggregates - Mailing list pgsql-performance

From Greg Stark
Subject Re: SLOW query with aggregates
Date
Msg-id 878yhqx65f.fsf@stark.xeocode.com
Whole thread Raw
In response to SLOW query with aggregates  ("A Palmblad" <adampalmblad@yahoo.ca>)
List pgsql-performance
"A Palmblad" <adampalmblad@yahoo.ca> writes:

>  GroupAggregate  (cost=0.00..338300.34 rows=884 width=345) (actual
> time=86943.272..382718.104 rows=3117 loops=1)
>    ->  Merge Join  (cost=0.00..93642.52 rows=1135610 width=345) (actual
> time=0.148..24006.748 rows=1120974 loops=1)

I think the reason you're getting a GroupAggregate here instead of a
HashAggregate is that the MergeJoin is already producing the records in the
desired order, so the GroupAggregate doesn't require an extra sort, ie, it's
effectively free.

You might be able to verify this by running the query with

enable_indexscan = off  and/or  enable_mergejoin = off

some combination of which might get the planner to do a seqscan of the large
table with a hash join to the small table and then a HashAggregate.

If you're reading a lot of the large table the seqscan could be a little
faster, not much though. And given the accurate statistics guesses here the
planner may well have gotten this one right and the seqscan is slower. Can't
hurt to be verify it though.

--
greg

pgsql-performance by date:

Previous
From: matt@ymogen.net
Date:
Subject: Re: [ADMIN] Benchmarking postgres on Solaris/Linux
Next
From: Mark Kirkwood
Date:
Subject: Re: Benchmarking postgres on Solaris/Linux