Re: large xml database - Mailing list pgsql-sql

From Viktor Bojović
Subject Re: large xml database
Date
Msg-id AANLkTinZqmQA5hovJNAdRoTTu3mjzfRQdw1=cmwVMFEk@mail.gmail.com
Whole thread Raw
In response to Re: large xml database  (James Cloos <cloos@jhcloos.com>)
Responses Re: large xml database  (Rob Sargent <robjsargent@gmail.com>)
List pgsql-sql


On Sun, Oct 31, 2010 at 2:26 AM, James Cloos <cloos@jhcloos.com> wrote:
>>>>> "VB" == Viktor Bojović <viktor.bojovic@gmail.com> writes:

VB> i have very big XML documment which is larger than 50GB and want to
VB> import it into databse, and transform it to relational schema.

Were I doing such a conversion, I'd use perl to convert the xml into
something which COPY can grok.  Any other language, script or compiled,
would work just as well.  The goal is to avoid having to slurp the whole
xml structure into memory.

-JimC
--
James Cloos <cloos@jhcloos.com>         OpenPGP: 1024D/ED7DAEA6

The insertion into dabase is not very big problem.
I insert it as XML docs, or as varchar lines or as XML docs in varchar format. Usually i use transaction and commit after block of 1000 inserts and it goes very fast. so insertion is over after few hours.
But the problem occurs when i want to transform it inside database from XML(varchar or XML format) into tables by parsing.
That processing takes too much time in database no matter if it is stored as varchar lines, varchar nodes or XML data type.

--
---------------------------------------
Viktor Bojović
---------------------------------------
Wherever I go, Murphy goes with me

pgsql-sql by date:

Previous
From: Viktor Bojović
Date:
Subject: Re: large xml database
Next
From: Rob Sargent
Date:
Subject: Re: large xml database