Re: Pushing PostgreSQL to the Limit (urgent!) - Mailing list pgsql-general

From Curt Sampson
Subject Re: Pushing PostgreSQL to the Limit (urgent!)
Date
Msg-id Pine.NEB.4.44.0207161808030.465-100000@angelic.cynic.net
Whole thread Raw
In response to Re: Pushing PostgreSQL to the Limit (urgent!)  (Chris Albertson <chrisalbertson90278@yahoo.com>)
Responses Re: Pushing PostgreSQL to the Limit (urgent!)
List pgsql-general
> --- Paulo Henrique Baptista de Oliveira
> <baptista@linuxsolutions.com.br> wrote:
>
> >     I will put it to insert 30 M (millions) registers by month (or 1
> > Million by day) in a year is about 400 Millions registers.
> >     Can pgsql support this? In What Machine?

Yes. A reasonably powerful PC with at least two nice fast IDE drives
should do the trick. I recommend you buy such a machine, set up
postgres, and start experimenting. It will probably take a couple
of weeks of work to figure out how to make your application run
efficiently.

On Mon, 15 Jul 2002, Chris Albertson wrote:
>
> I have a similar application.  I am storing astronomical data
> from a set of automated cameras.  The data just floods in
> forever.  I can see a billion rows in the future.
> I find that I _can_ keep up using only modest hardware IF I use
> "COPY" and not "INSERT" to input the data.  "COPY" is much, much
> faster.  Also indexes help with SELECT speed not really hurt
> COPY/INSERT speed so you need to ballance.

Right. You may find it worthwhile to drop the indexes, import, and rebuild
instead of import with the indexes in place, if you're not doing queries
at the same time. Or maybe partial indexes could help....

cjs
--
Curt Sampson  <cjs@cynic.net>   +81 90 7737 2974   http://www.netbsd.org
    Don't you know, in this new Dark Age, we're all light.  --XTC


pgsql-general by date:

Previous
From: Michael Meskes
Date:
Subject: Re: Embedded SQL in a function
Next
From: stefan@extum.com
Date:
Subject: sequence id