I am creating a web application that uses very small temporary tables
(i.e. tables that are created and then destroyed nearly immediately)
to store state data for a particular browser "session".
A table might have four columns and four rows each containing a small
piece of text.
The "sessions" last a minute at best.
I can contemplate (hope for?) as many as 100,000 sessions in an hour.
That would mean 100,000 table creations and destructions in an hour
( or getting on 2000 per minute).
Will that stress a postgres database? Stress defined as slow response
or exceeding storage.
What are the measures one would take to minimize stress?
I have looked at the documentation but don't find something exactly
apposite to this question.
Thanks,
John