| From: | Simon Riggs <simon(at)2ndQuadrant(dot)com> |
|---|---|
| To: | James David Smith <james(dot)david(dot)smith(at)gmail(dot)com> |
| Cc: | "pgsql-novice(at)postgresql(dot)org" <pgsql-novice(at)postgresql(dot)org> |
| Subject: | Re: General 'big data' advice.... |
| Date: | 2013-08-05 11:59:47 |
| Message-ID: | CA+U5nMK_OGqxeEGwgBusfYyumGigEuSHZqSVn4kgG5gRKDt3hg@mail.gmail.com |
| Views: | Whole Thread | Raw Message | Download mbox | Resend email |
| Thread: | |
| Lists: | pgsql-novice |
On 5 August 2013 12:38, James David Smith <james(dot)david(dot)smith(at)gmail(dot)com> wrote:
> Bit of an abstract question I appreciate, however I just thought I'd
> see what people thought. I have an anonymosied dataset of travel
> behaviour of some people in a major city (I'd rather not go into
> details if that's ok). What I intend to do is to work out where they
> each are for every minute of the day. So for ~80,000 people x 1440
> minutes = 115,200,000 rows of data! So a few questions:
>
> 1) Is PostgreSQL going to be able to cope with this? In terms of the
> table size? I think so...
>
> 2) My columns will be something like
> person_id integer,
> person_timestamp timestamp,
> person_location_geom geometry
> Any thoughts on those? The format of the columns?
>
> 3) I'll probably create a Primary Key which is a combination of
> person_id and person_timestamp. Does this sound like a good idea?
>
> 4) Should I use some indexes to improve performance maybe?
Try it and see. It really depends on the queries you will run.
--
Simon Riggs http://www.2ndQuadrant.com/
PostgreSQL Development, 24x7 Support, Training & Services
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Luca Ferrari | 2013-08-05 13:33:39 | Re: General 'big data' advice.... |
| Previous Message | James David Smith | 2013-08-05 11:38:13 | General 'big data' advice.... |