On Jun 19, 2014, at 11:21 AM, Andy Colson wrote:
> I think it depends on how you are going to use them. I, for example, have lots of images that are served on a web
page,after benchmarks I found it was faster to store them on filesystem and let apache serve them directly.
I rarely store images like that locally now; I just toss them onto Amazon S3.
When I did have to store lots of images locally , I found this to be the best method:
1. The Postgres record for the image is given a unique and random hash as a hexdigest
2. The Image is saved onto a filesystem into a directory mapped by the hexdigest
for example, there might be something like this:
Postgres:
id | filename | hash
001 | image.jpg | abcdef123
Filesystem
abc/def/123/abcdef123-image.jpg
nginx/apache rewrite rule :
abcdef123-image.jpg -> abc/def/123/abcdef123-image.jpg
the reason for this has to do with the performance of various filesystems and issues with the distribution of digits in
asequence. it ties into Benford's Law ( http://en.wikipedia.org/wiki/Benford's_law ) as well.
a handful of filesystems exhibit decreased performance as the number of items in a directory increases. a few years
ago,1k-4k items was a safe max -- but at 10x that some filesystems really slowed. i think most modern filesystems are
stillquick at the 5-10k range.
a hash has more characters and a more normal distribution than a series of numbers or natural language filenames.
and if you group a hexdigest into triplets , you get 4096 max files/folders in a directory which is a decent sweet
spot
16 * 16 * 16 = 4096
i haven't had to deal with this sort of stuff in almost 10 years now. but archiving content like this back then was a
considerableimprovement to filesystem performance and web serving.