Caching Websites - Mailing list pgsql-general

From Adam Kessel
Subject Caching Websites
Date
Msg-id 20030509204849.GC8583@bostoncoop.net
Whole thread Raw
Responses Re: Caching Websites  (Richard Huxton <dev@archonet.com>)
List pgsql-general
I'm writing a python script that (among other things) caches websites.
Ultimately, the data is all stored in a string (pickled, possibly
zipped).  (lots of related stuff in stored in postgresql tables).

I am wondering whether it would be better to store each website in a
record in a table, or instead have a table which links URLs to filenames
(the file would contain the pickled website).  The sites will of course
vary greatly in size, but typically be between 1k and 200k (I probably
won't store anything bigger than that).

This seems like a simple question, and I suspect there's an obvious
answer for which data storage method makes more sense, I just don't know
how to go about researching that.  What would be the considerations for
using one method of data storage vs. the other?

Any suggestions for me?

--Adam

Attachment

pgsql-general by date:

Previous
From: Keary Suska
Date:
Subject: semget: No space left on device
Next
From: Keary Suska
Date:
Subject: Re: semget: No space left on device