> My problem is, I need to benchmark set of tables, where - we can
> assume - schema of each table is unknown, and we have no assumption on
> any fields being present there. (altho, if there is no other way to do
> it, we could assume id bigint not null default nextval('someseq'));
> basically, I need to know when certain row was selected (read), and
> when it was updated/created (insert). For that I need two fields. And
> although former could be done, and I more or less know how to do it (I
> think it can be done with rule, and default = now()) - I have yet to
> find a way on how to do the same thing for select.
So basically you've got a slony replicated database that you want to
check (on a row by row level) when something gets read/inserted/deleted?
It seems like you would want to add three fields to each table:
last_read_time, last_update_time, and original_insert_time
Then you restrict all access to the table and use security definer
functions to allow access. These functions would also update said
metadata to the table. If you need an ongoing log of access to the
tables, you could always add an accessor log table that looked like:
User (postgres/MYUSER) / Action (Select/Update/Insert) / Column
(some_column_name) / New Value (blah-value)
Erm, that's if I understand your question right. :-/
-Mark