Thread: auto vacuum errors
HI Gurus
I'm seeing a issue with some temp tables in contrib_regression db , the auto vacuum job will not run for this tables due some files/QID's missing how can I fix this ?
The temp tables are growing too fast and they don't get empty/clean-up...any tips ? I 'm running PGSQL 8.3.6 on RHES.
Isabella
I'm seeing a issue with some temp tables in contrib_regression db , the auto vacuum job will not run for this tables due some files/QID's missing how can I fix this ?
The temp tables are growing too fast and they don't get empty/clean-up...any tips ? I 'm running PGSQL 8.3.6 on RHES.
Isabella
-- ----------------------------------------------------------- Isabella A. Ghiurea Isabella.Ghiurea@nrc-cnrc.gc.ca Canadian Astronomy Data Centre |http://www.nrc-cnrc.gc.ca/eng/services/hia/data-centre.html National Research Council of Canada, Herzberg Institute of Astrophysics 5071 West Saanich Road, Victoria BC V9E 2E7, Canada Phone: 250 363-3446 fax: 250 363-0045
Isabella Ghiurea <isabella.ghiurea@nrc-cnrc.gc.ca> writes: > I'm seeing a issue with some temp tables in contrib_regression db , the > auto vacuum job will not run for this tables due some files/QID's > missing how can I fix this ? Um ... in general, autovacuum just silently ignores temp tables, because it doesn't have any way to access their contents. (One of the reasons that temp tables are fast is that the backend owning a temp table keeps its data in local buffers, not in shared buffers that another process could access.) I wouldn't expect any "missing file" complaints though. What exactly were you seeing? regards, tom lane
Hi Tom , Here is my error , the missing files were removed by mistake , I have another similar db cfg with this tables:spherrtmp% and no issues. I 'm thinking to restore this tables content from this second db , will this eliminate the auto vacuum error ? 07:45:59.770 PDT,,,31112,,4ff5a8a7.7988,5,,2012-07-05 07:45:59 PDT,19/2479,0,ERROR,58P01,"could not open relation 26384/247389/248165: No such file or directory",,,,,"automatic vacuum of table ""contrib_regression.public.spheretmp5""",,, 2012-07-05 07:45:59.770 PDT,,,31112,,4ff5a8a7.7988,6,,2012-07-05 07:45:59 PDT,19/2481,0,ERROR,58P01,"could not open relation 26384/247389/248174: No such file or directory",,,,,"automatic vacuum of table ""contrib_regression.public.spheretmp6""",,, On 07/04/2012 09:18 PM, Tom Lane wrote: > Isabella Ghiurea<isabella.ghiurea@nrc-cnrc.gc.ca> writes: >> I'm seeing a issue with some temp tables in contrib_regression db , the >> auto vacuum job will not run for this tables due some files/QID's >> missing how can I fix this ? > Um ... in general, autovacuum just silently ignores temp tables, because > it doesn't have any way to access their contents. (One of the reasons > that temp tables are fast is that the backend owning a temp table keeps > its data in local buffers, not in shared buffers that another process > could access.) I wouldn't expect any "missing file" complaints though. > What exactly were you seeing? > > regards, tom lane -- ----------------------------------------------------------- Isabella A. Ghiurea Isabella.Ghiurea@nrc-cnrc.gc.ca Canadian Astronomy Data Centre |http://www.nrc-cnrc.gc.ca/eng/services/hia/data-centre.html National Research Council of Canada, Herzberg Institute of Astrophysics 5071 West Saanich Road, Victoria BC V9E 2E7, Canada Phone: 250 363-3446 fax: 250 363-0045
Isabella Ghiurea <isabella.ghiurea@nrc-cnrc.gc.ca> writes: > Here is my error , the missing files were removed by mistake , I have > another similar db cfg with this tables:spherrtmp% and no issues. > I 'm thinking to restore this tables content from this second db , > will this eliminate the auto vacuum error ? What you ought to do is drop the tables that are reported as causing the error. Or the whole database --- are you actually doing something useful in contrib_regression, or is that just left over from testing? regards, tom lane