RE: [JDBC] Possible large object bug? - Mailing list pgsql-bugs

From Joe Shevland
Subject RE: [JDBC] Possible large object bug?
Date
Msg-id HEECIHEEJDBMCCGMGIOBKEHCCAAA.shevlandj@kpi.com.au
Whole thread Raw
In response to Re: [JDBC] Possible large object bug?  (Peter Mount <peter@retep.org.uk>)
List pgsql-bugs
Hi,

Tom C. and I exchanged a few emails about the issue offlist; Tom spotted after a while that the PostgreSQL data had
beenclobbered by the output of my cron job (not possible in normal conditions I guess)... end result being he suspects
itsa hardware glitch or similar (also had 'wc' dump core that same night during the daily periodic run; I'm choosing
cosmicradiation ;). 

I'm going to upgrade to the latest release and also run up another shinier box and see if things work out, I'm sure
theywill (digging out a tripwire floppy to make ultra sure nothings changed). As a side issue, could/should the
vacuumlofunctionality be merged with vacuum? 

Cheers,
Joe

> -----Original Message-----
> From: pgsql-jdbc-owner@postgresql.org
> [mailto:pgsql-jdbc-owner@postgresql.org]On Behalf Of Peter Mount
> Sent: Thursday, March 29, 2001 6:52 PM
> To: Joe Shevland; pgsql-jdbc@postgresql.org
> Cc: pgsql-bugs@postgresql.org
> Subject: Re: [JDBC] Possible large object bug?
>
>
> At 10:37 27/03/01 +1000, Joe Shevland wrote:
> >Hi,
> >
> >Semi-topical I hope ;)
>
> Yes semi ;-)
>
> >  I've started using Postgres 7.1 (FreeBSD 4.2-S) and large objects via
> > JDBC. (postmaster (PostgreSQL) 7.1beta5)
>
> I'm forwarding this to the bugs list as it looks like something nasty in
> the back end.
>
>
> >Everything has been working nicely with storing/retrieving blobs, until
> >last night during a vacuum of the database the backend process crashed
> >with the messages added to the end of this email. I'm also using the
> >'vacuumlo' contributed code. The order of the cron jobs is:
> >
> >59 2  *  *  *  postgres /usr/local/pgsql/bin/vacuumlo -v db1 db2 db3
> >59 3  *  *  *  postgres /usr/local/pgsql/bin/vacuumdb -z db1
> >59 4  *  *  *  postgres /usr/local/pgsql/bin/vacuumdb -z db2
> >59 5  *  *  *  postgres /usr/local/pgsql/bin/vacuumdb -z db3
> >
> >so I was wondering if there might be a bug in the vacuumlo code (though
> >its vacuumdb dying)? Or I was thinking, because they're
> development db's,
> >that frequent dropping/recreating of tables is maybe causing the
> prob? The
> >same vacuum commands have run fine before, both from cron and
> the command
> >line, the only difference was slightly heavier
> dropping/recreating yesterday.
> >
> >I'm yet to see if that particular database is stuffed as I can recreate
> >and retest easily enough. Let me know if I can give any further info,
> >
> >Regards,
> >Joe
> >
> >---
> >NOTICE:  Rel pg_attribute: TID 1/115: OID IS INVALID. TUPGONE 1.
> >...
> >NOTICE:  Rel pg_attribute: TID 1/6087: OID IS INVALID. TUPGONE 1.
> >NOTICE:  Rel pg_attribute: TID 1/6111: OID IS INVALID. TUPGONE 1.
> >NOTICE:  Rel pg_attribute: TID 1/6112: OID IS INVALID. TUPGONE 1.
> >NOTICE:  Rel pg_attribute: TID 1/6136: OID IS INVALID. TUPGONE 1.
> >NOTICE:  Rel pg_attribute: TID 1/6137: OID IS INVALID. TUPGONE 1.
> >pqReadData() -- backend closed the channel unexpectedly.
> >         This probably means the backend terminated abnormally
> >         before or while processing the request.
> >connection to server was lost
> >vacuumdb: vacuum  db2 failed
> >---
> >
> >with ~500 of the NOTICE lines then the crash. About 1% give a TUPGONE 0
> >ending instead.
> >
> >
> >---------------------------(end of broadcast)---------------------------
> >TIP 1: subscribe and unsubscribe commands go to majordomo@postgresql.org
>
>
> ---------------------------(end of broadcast)---------------------------
> TIP 4: Don't 'kill -9' the postmaster
>


pgsql-bugs by date:

Previous
From: Peter Eisentraut
Date:
Subject: Re: unix_socket_directory isn't used by all bin tools
Next
From: sasha
Date:
Subject: PORTS : Windows 95/98/2000