> We have a PostgreSQL database full of over 500,000 events.
> When the database has a small number of events, to
> retrieve a specific event stored in that db takes a mere
> few seconds.
>
> However, as the database grew in size to where it is now,
> it takes over 15-20 seconds to get information back from a
> query to that database. We have tried everything,
> including vacuum, which someone else recommended to me.
Sometimes it's the simple things. Did you do an explain on the query to
see its access plan? I see nothing in your post about indexes.
> Am I correct in saying that 500,000 events shouldn't be a
> problem? Is there anyone out there with more than 500,000
> events in their DB than what we currently have?
With proper indexes, 500,000,000 "events" shouldn't be a problem.
--
pv
If this helped you, http://svcs.affero.net/rm.php?r=philovivero