Realistic upper bounds on DB size - Mailing list pgsql-general

From A.M.
Subject Realistic upper bounds on DB size
Date
Msg-id F03A2710-607A-11D7-9333-0030657192DA@cmu.edu
Whole thread Raw
List pgsql-general
I'm trying to figure out what the upper bounds on a postgresql table
are based on required use of indices and integer columns in a single
table.
    An astronomy institution I'm considering working for receives a
monster amount of telescope data from a government observatory. Each
day, they download millions of rows of data (including position in the
sky, infrared reading, etc.) in CSV format. Most of the rows are floats
and integers. I would like to offer them an improvement over their old
system.
    I would like to know how postgresql does under such extreme
circumstances- for example, I may load the entire millions of rows CSV
file into a table and then eliminate some odd million rows they are not
interested in. Would a vacuum at this time be prohibitively expensive?
If I add some odd millions of rows to a table every day, can I expect
the necessary indices to keep up? In other words, will postgresql be
able to keep up with their simple and infrequent selects on monster
amounts of data (potentially 15 GB/day moving in and out daily with db
growth at ~+5 GB/day [millions of rows] in big blocks all at once)
assuming that they have top-of-the-line equipment for this sort of
thing (storage, memory, processors, etc.)? Anyone else using postgresql
on heavy-duty astronomy data? Thanks for any info.
 ><><><><><><><><><
AgentM
agentm@cmu.edu


pgsql-general by date:

Previous
From: Ken Guest
Date:
Subject: Re: Reverse engineering PG database
Next
From: "Moritz Lennert"
Date:
Subject: Re: vacuum full fills up disk - solved