Re: table versioning approach (not auditing) - Mailing list pgsql-general

From Felix Kunde
Subject Re: table versioning approach (not auditing)
Date
Msg-id trinity-4c5f6cba-04af-4826-ab45-e9d38651cdac-1412155137210@3capp-gmx-bs33
Whole thread Raw
In response to Re: table versioning approach (not auditing)  (Adam Brusselback <adambrusselback@gmail.com>)
Responses Re: table versioning approach (not auditing)
List pgsql-general
Hey there. Thank you very much for that fix! Thats why I'd like to have a joint development and joint testing. It's way
moreconvincing for users to go for a solution that is tested by some experts than just by a random developer :)  

I'm open to create a new project and push the code there. Don't care about the name. Then we might figure out which
partsare already good, which parts could be improved and where to go next. I think switching to JSONB for example will
beeasy, as it offers the same functions than JSON afaik. 
 

Gesendet: Dienstag, 30. September 2014 um 21:16 Uhr
Von: "Adam Brusselback" <adambrusselback@gmail.com>
An: "Felix Kunde" <felix-kunde@gmx.de>
Cc: "pgsql-general@postgresql.org" <pgsql-general@postgresql.org>
Betreff: Re: [GENERAL] table versioning approach (not auditing)

Felix, I'd love to see a single, well maintained project. For example, I just found yours, and gave it a shot today
afterseeing this post.  I found a bug when an update command is issued, but the old and new values are all the same. 
Thetrigger will blow up.  I've got a fix for that, but if we had one project that more than a handful of people used,
stufflike that would be quashed very quickly. 
 
I love the design of it by the way. Any idea what it will take to move to JSONB for 9.4? 
 
 
On Tue, Sep 30, 2014 at 7:22 AM, Felix Kunde <felix-kunde@gmx.de> wrote:Hey
 
yes i'm adding an additional key to each of my tables. First i wanted to use the primary key as one column in my
audit_logtable, but in some of my tables the PK consists of more than one column. Plus it's nice to have one key that
iscalled the same over all tables. 
 
To get a former state for one row at date x I need to join the latest delta BEFORE date x with each delta AFTER date x.
IfI would log complete rows, this joining part would not be neccessary, but as I usually work with spatial databases
thathave complex geometries and also image files, this strategy is too harddisk consuming. 
 
If there are more users following a similar approach, I wonder why we not throw all the good ideas together, to have
onesolution that is tested, maintained and improved by more developpers. This would be great. 
 
Felix
 

Gesendet: Montag, 29. September 2014 um 23:25 Uhr
Von: "Abelard Hoffman" <abelardhoffman@gmail.com[abelardhoffman@gmail.com]>
An: "Felix Kunde" <felix-kunde@gmx.de[felix-kunde@gmx.de]>
Cc: "pgsql-general@postgresql.org[pgsql-general@postgresql.org]"
<pgsql-general@postgresql.org[pgsql-general@postgresql.org]>
Betreff: Re: [GENERAL] table versioning approach (not auditing)

Thank you Felix, Gavin, and Jonathan for your responses.
 
Felix & Jonathan: both of you mention just storing deltas. But if you do that, how do you associate the delta record
withthe original row? Where's the PK stored, if it wasn't part of the delta? 
 
Felix, thank you very much for the example code. I took a look at your table schemas. I need to study it more, but it
lookslike the way you're handling the PK, is you're adding a separate synthethic key (audit_id) to each table that's
beingversioned. And then storing that key along with the delta. 
 
So then to find all the versions of a given row, you just need to join the audit row with the
schema_name.table_name.audit_idcolumn. Is that right? The only potential drawback there is there's no referential
integritybetween the audit_log.audit_id and the actual table. 
 
I do like that approach very much though, in that it eliminates the need to interrogate the json data in order to
performmost queries. 
 
AH
 
 
 
On Mon, Sep 29, 2014 at 12:26 AM, Felix Kunde <felix-kunde@gmx.de[felix-kunde@gmx.de]> wrote:Hey
 
i've also tried to implement a database versioning using JSON to log changes in tables. Here it is:
https://github.com/fxku/audit[https://github.com/fxku/audit][https://github.com/fxku/audit[https://github.com/fxku/audit]]
I've got two versioning tables, one storing information about all transactions that happened and one where i put the
JSONlogs of row changes of each table. I'm only logging old values and not complete rows. 
 
Then I got a function that recreates a database state at a given time into a separate schema - either to VIEWs, MVIEWs
orTABLES. This database state could then be indexed in order to work with it. You can also reset the production state
tothe recreated past state. 
 
Unfortunately I've got no time to further work on it at the moment + I have not done tests with many changes in the
databaseso I can't say if the recreation process scales well. On downside I've realised is that using the json_agg
functionhas limits when I've got binary data. It gets too long. So I'm really looking forward using JSONB. 

There are more plans in my mind. By having a Transaction_Log table it should be possible to revert only certain
transactions.I'm also thinking of parallel versioning, e.g. different users are all working with their version of the
databaseand commit their changes to the production state. As I've got a unique history ID for each table and each row,
Ishould be able to map the affected records. 

Have a look and tell me what you think of it.

Cheers
Felix
 

Gesendet: Montag, 29. September 2014 um 04:00 Uhr
Von: "Abelard Hoffman" <abelardhoffman@gmail.com[abelardhoffman@gmail.com]>
An: "pgsql-general@postgresql.org[pgsql-general@postgresql.org]"
<pgsql-general@postgresql.org[pgsql-general@postgresql.org]>
Betreff: [GENERAL] table versioning approach (not auditing)

Hi. I need to maintain a record of all changes to certain tables so assist in viewing history and reverting changes
whennecessary (customer service makes an incorrect edit, etc.). 
 
I have studied these two audit trigger examples:

https://wiki.postgresql.org/wiki/Audit_trigger[https://wiki.postgresql.org/wiki/Audit_trigger][https://wiki.postgresql.org/wiki/Audit_trigger[https://wiki.postgresql.org/wiki/Audit_trigger]]
https://wiki.postgresql.org/wiki/Audit_trigger_91plus
 
I've also read about two other approaches to versioning:
1. maintain all versions in one table, with a flag to indicate which is the current version
2. have a separate versions table for each real table, and insert into the associated version table whenever an update
orinsert is done. 
 
My current implementation is based on the wiki trigger examples, using a single table, and a json column to record the
rowchanges (rather than hstore). What I like about that, in particular, is I can have a "global," chronological view of
allversioned changes very easily. 
 
But there are two types of queries I need to run.
1. Find all changes made by a specific user
2. Find all changes related to a specific record
 
#1 is simple to do. The versioning table has a user_id column of who made the change, so I can query on that.
 
#2 is more difficult. I may want to fetch all changes to a group of tables that are all related by foreign keys (e.g.,
findall changes to "user" record 849, along with any changes to their "articles," "photos," etc.). All of the data is
inthe json column, of course, but it seems like a pain to try and build a query on the json column that can fetch all
thoserelationships (and if I mess it up, I probably won't generate any errors, since the json is so free-form). 
 
So my question is, do you think using the json approach is wrong for this case? Does it seem better to have separate
versioningtables associated with each real table? Or another approach? 
 
Thanks
 
 

 

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org[pgsql-general@postgresql.org])
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general[http://www.postgresql.org/mailpref/pgsql-general]


pgsql-general by date:

Previous
From: Misa Simic
Date:
Subject: Re: improving speed of query that uses a multi-column "filter" ?
Next
From: Emanuel Araújo
Date:
Subject: Extract especific text from a sql statement