Sent from my iPad
On 18-Feb-2013, at 22:27, Bruce Momjian <bruce@momjian.us> wrote:
> On Mon, Feb 18, 2013 at 09:56:22AM -0600, Merlin Moncure wrote:
>> On Mon, Feb 18, 2013 at 5:10 AM, Seref Arikan
>> <serefarikan@kurumsalteknoloji.com> wrote:
>>> Greetings,
>>> What would happen if I create multiple threads from within a postgresql
>>> function written in C?
>>> I have the opportunity to do parallel processing on binary data, and I n=
eed
>>> to create multiple threads to do that.
>>> If I can ensure that all my threads complete their work before I exit my=
>>> function, would this cause any trouble ?
>>> I am aware of postgresql's single threaded nature when executing queries=
,
>>> but is this a limitation for custom multi threaded code use in C based
>>> functions?
>>> I can't see any problems other than my custom spawn threads living beyon=
d my
>>> function's execution and memory/resource allocation issues, but if I can=
>>> handle them, should not I be safe?
>>>=20
>>> I believe I've seen someone applying a similar principle to use GPUs wit=
h
>>> postgresql, and I'm quite interested in giving this a try, unless I'm
>>> missing something.
>>=20
>> Some things immediately jump to mind:
>> *) backend library routines are not multi-thread safe. Notably, the
>> SPI interface and the memory allocator, but potentially anything. So
>> your spawned threads should avoid calling the backend API. I don't
>> even know if it's safe to call malloc.
>>=20
>> *) postgres exception handling can burn you, so I'd be stricter than
>> "before I exit my function"...really, you need to make sure threads
>> terminate before any potentially exception throwing backend routine
>> fires, which is basically all of them including palloc memory
>> allocation and interrupt checking. So, we must understand that:
>>=20
>> While your threads are executing, your query can't be cancelled --
>> only a hard kill will take the database down. If you're ok with that
>> risk, then go for it. If you're not, then I'd thinking about
>> sendinging the bytea through a protocol to a threaded processing
>> server running outside of the database. More work and slower
>> (protocol overhead), but much more robust.
>=20
> You can see the approach of not calling any PG-specific routines from
> theads here:
>=20
> http://wiki.postgresql.org/wiki/Parallel_Query_Execution#Approaches
>=20
Is there any way to locally synchronise the threads in my code,and send the r=
equests to the PostgreSQL backend one at a time? Like a waiting queue in my c=
ode?
Regards,
Atri=