Re: splitting data into multiple tables - Mailing list pgsql-performance

From nair rajiv
Subject Re: splitting data into multiple tables
Date
Msg-id d67ff5e61001260717h6b049a3ch529b9894fb8e123a@mail.gmail.com
Whole thread Raw
In response to Re: splitting data into multiple tables  (Matthew Wakeling <matthew@flymine.org>)
List pgsql-performance


On Tue, Jan 26, 2010 at 5:15 PM, Matthew Wakeling <matthew@flymine.org> wrote:
On Mon, 25 Jan 2010, nair rajiv wrote:
I am working on a project that will take out structured content from wikipedia and put it in our database...

there is a table which will approximately have 5 crore entries after data
harvesting.

Have you asked the Wikimedia Foundation if they mind you consuming that much of their bandwidth, or even if there are copyright issues involved in grabbing that much of their data?

 
We are downloading the nt and owl files kept for download at
http://wiki.dbpedia.org/Downloads34


(The other problem with using the word "crore" is that although it may mean 10000000 in a few countries, it could also mean 500000.)

Matthew

--
Of course it's your fault. Everything here's your fault - it says so in your
contract.                                    - Quark

pgsql-performance by date:

Previous
From: Matthew Wakeling
Date:
Subject: Re: splitting data into multiple tables
Next
From: Mark Hills
Date:
Subject: Poor query plan across OR operator