> Pavel, thank you for your interest. > It is definitely something inside this document. > > Actually we loaded about 10k different documents like this one. About 10Gb > of content and crash is only on this one.
It's probably a good idea to report this to libxml2 then.
Sure, but bug is mostly about unhandled server crash. Is it normal?
Nobody can handle process crash. This should be fixed on libxml2 side.
Also, as far as I understand Pavel, lbxml2 interface somehow changed for handling 'huge' documents, but postgres don't handle this correctly.
This can be fixed only partially - the authors of libxml2 introduced new limit and new option, but there is not possible apply new option on all API.
If I understand well to this behave - libxml2 implemented new limits as safeguards against strange attacker documents. Is it question if it is good idea disable this safeguards by default. Second issue is impossibility to set this option for functionality that we use from libxml2 - and there are not alternatives. I have not any idea, how these issues can be fixed on postgresql side.
The mentioned limit is 10 000 000 bytes - if you will generate documents less than this size, then probably don't hit this issue.