From: | Adrian Klaver <adrian(dot)klaver(at)aklaver(dot)com> |
---|---|
To: | Edson Richter <edsonrichter(at)hotmail(dot)com>, pgsql-general(at)postgresql(dot)org |
Subject: | Re: Fastest way to duplicate a quite large database |
Date: | 2016-04-12 15:04:16 |
Message-ID: | 570D0E70.5050208@aklaver.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
On 04/12/2016 07:51 AM, Edson Richter wrote:
> Same machine, same cluster - just different database name.
Hmm, running tests against the same cluster you are running the
production database would seem to be a performance hit against the
production database and potentially dangerous should the tests trip a
bug that crashes the server.
>
> Atenciosamente,
>
> Edson Carlos Ericksson Richter
>
> Em 12/04/2016 11:46, John R Pierce escreveu:
>> On 4/12/2016 7:25 AM, Edson Richter wrote:
>>>
>>> I have a database "Customer" with about 60Gb of data.
>>> I know I can backup and restore, but this seems too slow.
>>>
>>> Is there any other option to duplicate this database as
>>> "CustomerTest" as fast as possible (even fastar than backup/restore)
>>> - better if in one operation (something like "copy database A to B")?
>>> I would like to run this everyday, overnight, with minimal impact to
>>> prepare a test environment based on production data.
>>
>>
>> copy to the same machine, or copy to a different test server?
>> different answers.
>>
>>
>>
>
>
>
--
Adrian Klaver
adrian(dot)klaver(at)aklaver(dot)com
From | Date | Subject | |
---|---|---|---|
Next Message | John R Pierce | 2016-04-12 15:14:51 | Re: Fastest way to duplicate a quite large database |
Previous Message | Louis Battuello | 2016-04-12 15:03:34 | Re: Fastest way to duplicate a quite large database |