An interesting suggestion, but the problem with storing the logfiles in
a table for us would be that uncompressed, the log file for a day is
about 10 gigs. This would mean that an unacceptable amount of excess
data would accumulate in the database. It would be feasible however to
write a script to import the archived logfile into a new temporary
database on a different server, then use SQL to search it and delete the
db when finished.
Thanks for the suggestion though.
Ben.
Alvaro Herrera wrote:
> Ben Webber wrote:
>> Hi,
>>
>> I wrote a shell script to find the duration and the related statement in the log file and place them one after the
otherif the duration is over a specified time like this:-
>>
>> 2008-10-31 02:00:49 GMT [23683] [mp_live] LOG: statement: CLUSTER;
>> 2008-10-31 02:04:42 GMT [23683] [mp_live] LOG: duration: 232783.684 ms
>
> I wonder if you'd benefit from doing CSV logs and then storing them into
> a table. Querying using SQL is probably going to be easier (and more
> robust -- it'd work even with embedded newlines etc).
>
This message has been scanned for malware by SurfControl plc. www.surfcontrol.com