Re: Thoughts on how to avoid a massive integer update. - Mailing list pgsql-general

From David G. Johnston
Subject Re: Thoughts on how to avoid a massive integer update.
Date
Msg-id CAKFQuwagTKMw0hkxhiwhvXZwto-sEJa1EF04e6vDaaaUcUgdsQ@mail.gmail.com
Whole thread Raw
In response to Thoughts on how to avoid a massive integer update.  ("Fehrle, Brian" <bfehrle@comscore.com>)
Responses Re: Thoughts on how to avoid a massive integer update.
List pgsql-general
On Mon, May 4, 2020 at 2:33 PM Fehrle, Brian <bfehrle@comscore.com> wrote:

I NEED to update every single row in all these tables, changing the integer value to a different integer.

 

Does anyone have any hackery ideas on how to achieve this in less time?


Probably the only solution that would perform computationally faster would take the same amount of time or more to code and debug, and be considerably riskier.  Basically shut down PostgreSQL and modify the data files directly to change one integer byte sequence to another.  On the positive side the source code for PostgreSQL is open source and that data, while complex, is structured.

On the point of "vacuum" versus "vacuum full" - I don't know if this is how it would work in reality but conceptually if you updated half the table, vacuumed, updated the second half, vacuumed, the second batch of updates would reuse the spaced freed from the first batch and you'd only increase the disk consumption by 1.5 instead of 2.0.  As you increase the number of batches the percentage of additional space consumed decreases.  Though if you have the space I'd have to imagine that creating a brand new table and dropping the old one would be the best solution when taken in isolation.

David J.

pgsql-general by date:

Previous
From: "David G. Johnston"
Date:
Subject: Re: Best way to use trigger to email a report ?
Next
From: "David G. Johnston"
Date:
Subject: Re: Best way to use trigger to email a report ?