> You might want to reduce random_page_cost a little.
> Keep in mind that your test case is small enough to fit in RAM and is
> probably not reflective of what will happen with larger tables.
I am also running 8.0 rc1 for Windows. Despite many hours spent tweaking various planner cost constants, I found little
effecton cost estimates. Even reducing random_page_cost from 4.0 to 0.1 had negligible impact and failed to
significantlyinfluence the planner.
Increasing the statistics target for the last_name column to 250 or so *may* help, at least if you're only selecting
onename at a time. That's the standard advice around here and the only thing I've found useful. Half the threads in
thisforum are about under-utilized indexes. It would be great if someone could admit the planner is broken and talk
aboutactually fixing it!
I'm unconvinced that the planner only favours sequential scans as table size decreases. In my experience so far, larger
tableshave the same problem only it's more noticeable.
The issue hits PostgreSQL harder than others because of its awful sequential scan speed, which is two to five times
slowerthan other DBMS. The archives show there has been talk for years about this, but it seems, no solution. The
obviousthing to consider is the block size, but people have tried increasing this in the past with only marginal
success.
Regards
David