From: | nair rajiv <nair331(at)gmail(dot)com> |
---|---|
To: | Matthew Wakeling <matthew(at)flymine(dot)org>, pgsql-performance(at)postgresql(dot)org |
Subject: | Re: splitting data into multiple tables |
Date: | 2010-01-26 15:17:09 |
Message-ID: | d67ff5e61001260717h6b049a3ch529b9894fb8e123a@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-performance |
On Tue, Jan 26, 2010 at 5:15 PM, Matthew Wakeling <matthew(at)flymine(dot)org>wrote:
> On Mon, 25 Jan 2010, nair rajiv wrote:
>
>> I am working on a project that will take out structured content from
>> wikipedia and put it in our database...
>>
>> there is a table which will approximately have 5 crore entries after data
>> harvesting.
>>
>
> Have you asked the Wikimedia Foundation if they mind you consuming that
> much of their bandwidth, or even if there are copyright issues involved in
> grabbing that much of their data?
>
We are downloading the nt and owl files kept for download at
http://wiki.dbpedia.org/Downloads34
> (The other problem with using the word "crore" is that although it may mean
> 10000000 in a few countries, it could also mean 500000.)
>
> Matthew
>
> --
> Of course it's your fault. Everything here's your fault - it says so in
> your
> contract. - Quark
>
From | Date | Subject | |
---|---|---|---|
Next Message | Mark Hills | 2010-01-26 16:00:40 | Poor query plan across OR operator |
Previous Message | Matthew Wakeling | 2010-01-26 11:45:41 | Re: splitting data into multiple tables |