Thread: template1 database is facing high datfrozenxid
Hi,
I am running Postgresql 9.1 and I can see the datfrozenxid is going high and vacuum process is not bringing it down. And this has been happening on template1 database.
2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING: oldest xmin is far in the past
2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT: Close open transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic vacuum of table "template1.pg_catalog.pg_database": index scans: 0
pages: 0 removed, 1 remain
tuples: 0 removed, 9 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING: oldest xmin is far in the past
2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT: Close open transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic vacuum of table "template1.pg_catalog.pg_largeobject": index scans: 0
pages: 0 removed, 0 remain
tuples: 0 removed, 0 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
I vacuum database manually but nothing is working out.
Please help,
Thanks.
On 02/12/2016 02:56 PM, AI Rumman wrote: > Hi, > > I am running Postgresql 9.1 and I can see the datfrozenxid is going high > and vacuum process is not bringing it down. And this has been happening > on template1 database. > > 2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING: oldest > xmin is far in the past > 2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT: Close open > transactions soon to avoid wraparound problems. The above seems to be the contributing factor. Does: select * from pg_stat_activity show long running queries. > 2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_database": index scans: 0 > pages: 0 removed, 1 remain > tuples: 0 removed, 9 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > 2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING: oldest > xmin is far in the past > 2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT: Close open > transactions soon to avoid wraparound problems. > 2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_largeobject": index scans: 0 > pages: 0 removed, 0 remain > tuples: 0 removed, 0 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > > > I vacuum database manually but nothing is working out. > Please help, > > Thanks. -- Adrian Klaver adrian.klaver@aklaver.com
I checked it and I did not find any log running sql or any open transaction. Not even in pg_prepared_xacts.
And it looks like pg_catalog database is making the alarm.
Any other idea please, where I need to look into.
Thanks.
On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaver <adrian.klaver@aklaver.com> wrote:
On 02/12/2016 02:56 PM, AI Rumman wrote:Hi,
I am running Postgresql 9.1 and I can see the datfrozenxid is going high
and vacuum process is not bringing it down. And this has been happening
on template1 database.
2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING: oldest
xmin is far in the past
2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT: Close open
transactions soon to avoid wraparound problems.
The above seems to be the contributing factor.
Does:
select * from pg_stat_activity
show long running queries.--2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_database": index scans: 0
pages: 0 removed, 1 remain
tuples: 0 removed, 9 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING: oldest
xmin is far in the past
2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT: Close open
transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_largeobject": index scans: 0
pages: 0 removed, 0 remain
tuples: 0 removed, 0 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
I vacuum database manually but nothing is working out.
Please help,
Thanks.
Adrian Klaver
adrian.klaver@aklaver.com
On 02/12/2016 03:10 PM, AI Rumman wrote: > I checked it and I did not find any log running sql or any open > transaction. Not even in pg_prepared_xacts. > And it looks like pg_catalog database is making the alarm. > > Any other idea please, where I need to look into. select * from pg_locks http://www.postgresql.org/docs/9.1/static/view-pg-locks.html > > Thanks. > > > On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaver > <adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote: > > On 02/12/2016 02:56 PM, AI Rumman wrote: > > Hi, > > I am running Postgresql 9.1 and I can see the datfrozenxid is > going high > and vacuum process is not bringing it down. And this has been > happening > on template1 database. > > 2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT: > Close open > transactions soon to avoid wraparound problems. > > > The above seems to be the contributing factor. > > Does: > > select * from pg_stat_activity > > show long running queries. > > > 2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_database": index > scans: 0 > pages: 0 removed, 1 remain > tuples: 0 removed, 9 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > 2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT: > Close open > transactions soon to avoid wraparound problems. > 2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_largeobject": > index scans: 0 > pages: 0 removed, 0 remain > tuples: 0 removed, 0 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > > > I vacuum database manually but nothing is working out. > Please help, > > Thanks. > > > > -- > Adrian Klaver > adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com> > > -- Adrian Klaver adrian.klaver@aklaver.com
On 02/12/2016 03:10 PM, AI Rumman wrote: > I checked it and I did not find any log running sql or any open > transaction. Not even in pg_prepared_xacts. > And it looks like pg_catalog database is making the alarm. > > Any other idea please, where I need to look into. Should have added: select * from pg_database > > Thanks. > > > On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaver > <adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote: > > On 02/12/2016 02:56 PM, AI Rumman wrote: > > Hi, > > I am running Postgresql 9.1 and I can see the datfrozenxid is > going high > and vacuum process is not bringing it down. And this has been > happening > on template1 database. > > 2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT: > Close open > transactions soon to avoid wraparound problems. > > > The above seems to be the contributing factor. > > Does: > > select * from pg_stat_activity > > show long running queries. > > > 2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_database": index > scans: 0 > pages: 0 removed, 1 remain > tuples: 0 removed, 9 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > 2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT: > Close open > transactions soon to avoid wraparound problems. > 2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic > vacuum of table "template1.pg_catalog.pg_largeobject": > index scans: 0 > pages: 0 removed, 0 remain > tuples: 0 removed, 0 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > > > I vacuum database manually but nothing is working out. > Please help, > > Thanks. > > > > -- > Adrian Klaver > adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com> > > -- Adrian Klaver adrian.klaver@aklaver.com
Used this query in each of the database::
SELECT t.relname, l.database, l.locktype, l.pid , l.mode, l.granted, p.current_query, p.query_start ,p.waiting
FROM pg_locks as l
INNER JOIN pg_stat_all_tables t
on l.relation = t.relid
INNER JOIN pg_stat_activity as p
on l.pid = p.procpid ;
No luck. At present, db is working, but t is going towards wraparound.
On Fri, Feb 12, 2016 at 3:28 PM, Adrian Klaver <adrian.klaver@aklaver.com> wrote:
On 02/12/2016 03:10 PM, AI Rumman wrote:I checked it and I did not find any log running sql or any open
transaction. Not even in pg_prepared_xacts.
And it looks like pg_catalog database is making the alarm.
Any other idea please, where I need to look into.
Should have added:
select * from pg_database
Thanks.
On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaveradrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com><adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote:
On 02/12/2016 02:56 PM, AI Rumman wrote:
Hi,
I am running Postgresql 9.1 and I can see the datfrozenxid is
going high
and vacuum process is not bringing it down. And this has been
happening
on template1 database.
2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT:
Close open
transactions soon to avoid wraparound problems.
The above seems to be the contributing factor.
Does:
select * from pg_stat_activity
show long running queries.
2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_database": index
scans: 0
pages: 0 removed, 1 remain
tuples: 0 removed, 9 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT:
Close open
transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_largeobject":
index scans: 0
pages: 0 removed, 0 remain
tuples: 0 removed, 0 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
I vacuum database manually but nothing is working out.
Please help,
Thanks.
--
Adrian Klaver
--
Adrian Klaver
adrian.klaver@aklaver.com
In pg_subtrans, I have files like:
$ ls -lrt | more
total 1269436
-rw------- 1 postgres postgres 262144 Jan 25 18:49 D907
-rw------- 1 postgres postgres 262144 Jan 25 18:54 D908
-rw------- 1 postgres postgres 262144 Jan 25 18:58 D909
-rw------- 1 postgres postgres 262144 Jan 25 18:59 D90A
-rw------- 1 postgres postgres 262144 Jan 25 19:04 D90B
-rw------- 1 postgres postgres 262144 Jan 25 19:09 D90C
-rw------- 1 postgres postgres 262144 Jan 25 19:14 D90D
-rw------- 1 postgres postgres 262144 Jan 25 19:18 D90E
-rw------- 1 postgres postgres 262144 Jan 25 19:19 D90F
-rw------- 1 postgres postgres 262144 Jan 25 19:24 D910
-rw------- 1 postgres postgres 262144 Jan 25 19:29 D911
-rw------- 1 postgres postgres 262144 Jan 25 19:33 D912
-rw------- 1 postgres postgres 262144 Jan 25 19:34 D913
-rw------- 1 postgres postgres 262144 Jan 25 19:39 D914
-rw------- 1 postgres postgres 262144 Jan 25 19:44 D915
-rw------- 1 postgres postgres 262144 Jan 25 19:49 D916
-rw------- 1 postgres postgres 262144 Jan 25 19:53 D917
-rw------- 1 postgres postgres 262144 Jan 25 19:54 D918
-rw------- 1 postgres postgres 262144 Jan 25 19:59 D919
-rw------- 1 postgres postgres 262144 Jan 25 20:04 D91A
-rw------- 1 postgres postgres 262144 Jan 25 20:09 D91B
-rw------- 1 postgres postgres 262144 Jan 25 20:14 D91C
-rw------- 1 postgres postgres 262144 Jan 25 20:19 D91D
-rw------- 1 postgres postgres 262144 Jan 25 20:23 D91E
-rw------- 1 postgres postgres 262144 Jan 25 20:24 D91F
-rw------- 1 postgres postgres 262144 Jan 25 20:29 D920
-rw------- 1 postgres postgres 262144 Jan 25 20:34 D921
-rw------- 1 postgres postgres 262144 Jan 25 20:39 D922
-rw------- 1 postgres postgres 262144 Jan 25 20:44 D923
-rw------- 1 postgres postgres 262144 Jan 25 20:49 D924
-rw------- 1 postgres postgres 262144 Jan 25 20:54 D925
-rw------- 1 postgres postgres 262144 Jan 25 20:59 D926
-rw------- 1 postgres postgres 262144 Jan 25 21:04 D927
.....
Does it mean that I have too many open transactions? If yes, it is not showing in pg_stat_activity.
On Fri, Feb 12, 2016 at 3:38 PM, AI Rumman <rummandba@gmail.com> wrote:
Used this query in each of the database::SELECT t.relname, l.database, l.locktype, l.pid , l.mode, l.granted, p.current_query, p.query_start ,p.waitingFROM pg_locks as lINNER JOIN pg_stat_all_tables ton l.relation = t.relidINNER JOIN pg_stat_activity as pon l.pid = p.procpid ;No luck. At present, db is working, but t is going towards wraparound.On Fri, Feb 12, 2016 at 3:28 PM, Adrian Klaver <adrian.klaver@aklaver.com> wrote:On 02/12/2016 03:10 PM, AI Rumman wrote:I checked it and I did not find any log running sql or any open
transaction. Not even in pg_prepared_xacts.
And it looks like pg_catalog database is making the alarm.
Any other idea please, where I need to look into.
Should have added:
select * from pg_database
Thanks.
On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaveradrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com><adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote:
On 02/12/2016 02:56 PM, AI Rumman wrote:
Hi,
I am running Postgresql 9.1 and I can see the datfrozenxid is
going high
and vacuum process is not bringing it down. And this has been
happening
on template1 database.
2016-02-12 16:51:50.400 CST [19445][@] : [13-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.400 CST [19445][@] : [14-1] HINT:
Close open
transactions soon to avoid wraparound problems.
The above seems to be the contributing factor.
Does:
select * from pg_stat_activity
show long running queries.
2016-02-12 16:51:50.400 CST [19445][@] : [15-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_database": index
scans: 0
pages: 0 removed, 1 remain
tuples: 0 removed, 9 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
2016-02-12 16:51:50.411 CST [19435][@] : [82-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.411 CST [19435][@] : [83-1] HINT:
Close open
transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.411 CST [19435][@] : [84-1] LOG: automatic
vacuum of table "template1.pg_catalog.pg_largeobject":
index scans: 0
pages: 0 removed, 0 remain
tuples: 0 removed, 0 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
I vacuum database manually but nothing is working out.
Please help,
Thanks.
--
Adrian Klaver
--
Adrian Klaver
adrian.klaver@aklaver.com
On 02/12/2016 04:03 PM, AI Rumman wrote: > In pg_subtrans, I have files like: Are you sure you are looking at the same database cluster in all the cases? What does: SELECT datname, age(datfrozenxid) FROM pg_database; give you? > > > $ ls -lrt | more > total 1269436 > -rw------- 1 postgres postgres 262144 Jan 25 18:49 D907 > -rw------- 1 postgres postgres 262144 Jan 25 18:54 D908 > -rw------- 1 postgres postgres 262144 Jan 25 18:58 D909 > -rw------- 1 postgres postgres 262144 Jan 25 18:59 D90A > -rw------- 1 postgres postgres 262144 Jan 25 19:04 D90B > -rw------- 1 postgres postgres 262144 Jan 25 19:09 D90C > -rw------- 1 postgres postgres 262144 Jan 25 19:14 D90D > -rw------- 1 postgres postgres 262144 Jan 25 19:18 D90E > -rw------- 1 postgres postgres 262144 Jan 25 19:19 D90F > -rw------- 1 postgres postgres 262144 Jan 25 19:24 D910 > -rw------- 1 postgres postgres 262144 Jan 25 19:29 D911 > -rw------- 1 postgres postgres 262144 Jan 25 19:33 D912 > -rw------- 1 postgres postgres 262144 Jan 25 19:34 D913 > -rw------- 1 postgres postgres 262144 Jan 25 19:39 D914 > -rw------- 1 postgres postgres 262144 Jan 25 19:44 D915 > -rw------- 1 postgres postgres 262144 Jan 25 19:49 D916 > -rw------- 1 postgres postgres 262144 Jan 25 19:53 D917 > -rw------- 1 postgres postgres 262144 Jan 25 19:54 D918 > -rw------- 1 postgres postgres 262144 Jan 25 19:59 D919 > -rw------- 1 postgres postgres 262144 Jan 25 20:04 D91A > -rw------- 1 postgres postgres 262144 Jan 25 20:09 D91B > -rw------- 1 postgres postgres 262144 Jan 25 20:14 D91C > -rw------- 1 postgres postgres 262144 Jan 25 20:19 D91D > -rw------- 1 postgres postgres 262144 Jan 25 20:23 D91E > -rw------- 1 postgres postgres 262144 Jan 25 20:24 D91F > -rw------- 1 postgres postgres 262144 Jan 25 20:29 D920 > -rw------- 1 postgres postgres 262144 Jan 25 20:34 D921 > -rw------- 1 postgres postgres 262144 Jan 25 20:39 D922 > -rw------- 1 postgres postgres 262144 Jan 25 20:44 D923 > -rw------- 1 postgres postgres 262144 Jan 25 20:49 D924 > -rw------- 1 postgres postgres 262144 Jan 25 20:54 D925 > -rw------- 1 postgres postgres 262144 Jan 25 20:59 D926 > -rw------- 1 postgres postgres 262144 Jan 25 21:04 D927 > ..... > > > Does it mean that I have too many open transactions? If yes, it is not > showing in pg_stat_activity. > > On Fri, Feb 12, 2016 at 3:38 PM, AI Rumman <rummandba@gmail.com > <mailto:rummandba@gmail.com>> wrote: > > Used this query in each of the database:: > > SELECT t.relname, l.database, l.locktype, l.pid , l.mode, l.granted, > p.current_query, p.query_start ,p.waiting > FROM pg_locks as l > INNER JOIN pg_stat_all_tables t > on l.relation = t.relid > INNER JOIN pg_stat_activity as p > on l.pid = p.procpid ; > > No luck. At present, db is working, but t is going towards wraparound. > > On Fri, Feb 12, 2016 at 3:28 PM, Adrian Klaver > <adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote: > > On 02/12/2016 03:10 PM, AI Rumman wrote: > > I checked it and I did not find any log running sql or any open > transaction. Not even in pg_prepared_xacts. > And it looks like pg_catalog database is making the alarm. > > Any other idea please, where I need to look into. > > > Should have added: > > select * from pg_database > > > Thanks. > > > On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaver > <adrian.klaver@aklaver.com > <mailto:adrian.klaver@aklaver.com> > <mailto:adrian.klaver@aklaver.com > <mailto:adrian.klaver@aklaver.com>>> wrote: > > On 02/12/2016 02:56 PM, AI Rumman wrote: > > Hi, > > I am running Postgresql 9.1 and I can see the > datfrozenxid is > going high > and vacuum process is not bringing it down. And > this has been > happening > on template1 database. > > 2016-02-12 16:51:50.400 CST [19445][@] : > [13-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.400 CST [19445][@] : > [14-1] HINT: > Close open > transactions soon to avoid wraparound problems. > > > The above seems to be the contributing factor. > > Does: > > select * from pg_stat_activity > > show long running queries. > > > 2016-02-12 16:51:50.400 CST [19445][@] : > [15-1] LOG: automatic > vacuum of table > "template1.pg_catalog.pg_database": index > scans: 0 > pages: 0 removed, 1 remain > tuples: 0 removed, 9 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > 2016-02-12 16:51:50.411 CST [19435][@] : > [82-1] WARNING: > oldest > xmin is far in the past > 2016-02-12 16:51:50.411 CST [19435][@] : > [83-1] HINT: > Close open > transactions soon to avoid wraparound problems. > 2016-02-12 16:51:50.411 CST [19435][@] : > [84-1] LOG: automatic > vacuum of table > "template1.pg_catalog.pg_largeobject": > index scans: 0 > pages: 0 removed, 0 remain > tuples: 0 removed, 0 remain > system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec > > > I vacuum database manually but nothing is working out. > Please help, > > Thanks. > > > > -- > Adrian Klaver > adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com> > <mailto:adrian.klaver@aklaver.com > <mailto:adrian.klaver@aklaver.com>> > > > > > -- > Adrian Klaver > adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com> > > > -- Adrian Klaver adrian.klaver@aklaver.com
Yes. And we solved the problem.
We looked into the pg_subtrans and found that we had subrans pending from January 25th. We investigated more and found that I large sql was executed on Streaming standby around that date.
More digging we found the date of the below alert is also near:
WARNING: oldest xmin is far in the past
We stopped standby and the problem solved. :)
Thanks.
On Fri, Feb 12, 2016 at 4:11 PM, Adrian Klaver <adrian.klaver@aklaver.com> wrote:
On 02/12/2016 04:03 PM, AI Rumman wrote:In pg_subtrans, I have files like:
Are you sure you are looking at the same database cluster in all the cases?
What does:
SELECT datname, age(datfrozenxid) FROM pg_database;
give you?<mailto:rummandba@gmail.com>> wrote:
$ ls -lrt | more
total 1269436
-rw------- 1 postgres postgres 262144 Jan 25 18:49 D907
-rw------- 1 postgres postgres 262144 Jan 25 18:54 D908
-rw------- 1 postgres postgres 262144 Jan 25 18:58 D909
-rw------- 1 postgres postgres 262144 Jan 25 18:59 D90A
-rw------- 1 postgres postgres 262144 Jan 25 19:04 D90B
-rw------- 1 postgres postgres 262144 Jan 25 19:09 D90C
-rw------- 1 postgres postgres 262144 Jan 25 19:14 D90D
-rw------- 1 postgres postgres 262144 Jan 25 19:18 D90E
-rw------- 1 postgres postgres 262144 Jan 25 19:19 D90F
-rw------- 1 postgres postgres 262144 Jan 25 19:24 D910
-rw------- 1 postgres postgres 262144 Jan 25 19:29 D911
-rw------- 1 postgres postgres 262144 Jan 25 19:33 D912
-rw------- 1 postgres postgres 262144 Jan 25 19:34 D913
-rw------- 1 postgres postgres 262144 Jan 25 19:39 D914
-rw------- 1 postgres postgres 262144 Jan 25 19:44 D915
-rw------- 1 postgres postgres 262144 Jan 25 19:49 D916
-rw------- 1 postgres postgres 262144 Jan 25 19:53 D917
-rw------- 1 postgres postgres 262144 Jan 25 19:54 D918
-rw------- 1 postgres postgres 262144 Jan 25 19:59 D919
-rw------- 1 postgres postgres 262144 Jan 25 20:04 D91A
-rw------- 1 postgres postgres 262144 Jan 25 20:09 D91B
-rw------- 1 postgres postgres 262144 Jan 25 20:14 D91C
-rw------- 1 postgres postgres 262144 Jan 25 20:19 D91D
-rw------- 1 postgres postgres 262144 Jan 25 20:23 D91E
-rw------- 1 postgres postgres 262144 Jan 25 20:24 D91F
-rw------- 1 postgres postgres 262144 Jan 25 20:29 D920
-rw------- 1 postgres postgres 262144 Jan 25 20:34 D921
-rw------- 1 postgres postgres 262144 Jan 25 20:39 D922
-rw------- 1 postgres postgres 262144 Jan 25 20:44 D923
-rw------- 1 postgres postgres 262144 Jan 25 20:49 D924
-rw------- 1 postgres postgres 262144 Jan 25 20:54 D925
-rw------- 1 postgres postgres 262144 Jan 25 20:59 D926
-rw------- 1 postgres postgres 262144 Jan 25 21:04 D927
.....
Does it mean that I have too many open transactions? If yes, it is not
showing in pg_stat_activity.
On Fri, Feb 12, 2016 at 3:38 PM, AI Rumman <rummandba@gmail.com
Used this query in each of the database::
SELECT t.relname, l.database, l.locktype, l.pid , l.mode, l.granted,
p.current_query, p.query_start ,p.waiting
FROM pg_locks as l
INNER JOIN pg_stat_all_tables t
on l.relation = t.relid
INNER JOIN pg_stat_activity as p
on l.pid = p.procpid ;
No luck. At present, db is working, but t is going towards wraparound.
On Fri, Feb 12, 2016 at 3:28 PM, Adrian Klaver
<adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>> wrote:
On 02/12/2016 03:10 PM, AI Rumman wrote:
I checked it and I did not find any log running sql or any open
transaction. Not even in pg_prepared_xacts.
And it looks like pg_catalog database is making the alarm.
Any other idea please, where I need to look into.
Should have added:
select * from pg_database
Thanks.
On Fri, Feb 12, 2016 at 3:05 PM, Adrian Klaver
<adrian.klaver@aklaver.com
<mailto:adrian.klaver@aklaver.com>
<mailto:adrian.klaver@aklaver.com<mailto:adrian.klaver@aklaver.com
<mailto:adrian.klaver@aklaver.com>>> wrote:
On 02/12/2016 02:56 PM, AI Rumman wrote:
Hi,
I am running Postgresql 9.1 and I can see the
datfrozenxid is
going high
and vacuum process is not bringing it down. And
this has been
happening
on template1 database.
2016-02-12 16:51:50.400 CST [19445][@] :
[13-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.400 CST [19445][@] :
[14-1] HINT:
Close open
transactions soon to avoid wraparound problems.
The above seems to be the contributing factor.
Does:
select * from pg_stat_activity
show long running queries.
2016-02-12 16:51:50.400 CST [19445][@] :
[15-1] LOG: automatic
vacuum of table
"template1.pg_catalog.pg_database": index
scans: 0
pages: 0 removed, 1 remain
tuples: 0 removed, 9 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
2016-02-12 16:51:50.411 CST [19435][@] :
[82-1] WARNING:
oldest
xmin is far in the past
2016-02-12 16:51:50.411 CST [19435][@] :
[83-1] HINT:
Close open
transactions soon to avoid wraparound problems.
2016-02-12 16:51:50.411 CST [19435][@] :
[84-1] LOG: automatic
vacuum of table
"template1.pg_catalog.pg_largeobject":
index scans: 0
pages: 0 removed, 0 remain
tuples: 0 removed, 0 remain
system usage: CPU 0.00s/0.00u sec elapsed 0.00 sec
I vacuum database manually but nothing is working out.
Please help,
Thanks.
--
Adrian Klaver
adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>
<mailto:adrian.klaver@aklaver.com>>
--
Adrian Klaver
adrian.klaver@aklaver.com <mailto:adrian.klaver@aklaver.com>
--
Adrian Klaver
adrian.klaver@aklaver.com