Re: PostGreSQL Replication and question on maintenance

2019-11-26 Thread github kran
Thanks Jeff for your reply.  We are not using replication concept here but
doing the data copy using few custom sql scripts. We had our PostgreSQL
engine supported by AWS and the instance was restarted automatically based
on the AWS design of PostgreSQL engine. Replica is restarted automatically
when the data sync is behind the master.

My intension is to bring to this forum and see if this is expected based on
the use case what I mentioned. I love this community and advices the
community has for various problems.  i still need to find out how can I do
this using replication. Please provide me few resources if there is a
better solution for this problem.

Thanks
githubkran



On Sat, Nov 16, 2019 at 11:13 AM Jeff Janes  wrote:

> On Thu, Nov 14, 2019 at 12:23 PM github kran  wrote:
>
>>
>>>
>>> *Problem what we have right now. *
>>>
>>> When the migration activity runs(weekly) from past 2 times , we saw the
>>> cluster read replica instance has restarted as it fallen behind the
>>> master(writer instance).
>>>
>>
> I can't figure out what your setup is here.  You must be using logical
> replication (not physical) or you wouldn't be able to write to the replica
> at all.  But if you are using logical replication, why do you also need
> these weekly jobs?  Why isn't logical replication taking care of it?
>
>
>
>
>> Everything
>>>
>>> after that worked seamlessly but we want to avoid the replica getting
>>> restarted. To avoid from restart we started doing smaller binary files and
>>> copy those files to the cluster-2
>>>
>>
> Who restarted it?  I am not aware of any case where the replica responds
> to falling behind by restarting itself.  With physical replication, it can
> start cancelling queries, but you don't seem to be using physical
> replication.
>
> Cheers,
>
> Jeff
>
>>


Re: PostGreSQL Replication and question on maintenance

2019-11-20 Thread Pavel Stehule
Hi

st 20. 11. 2019 v 13:12 odesílatel Soto Cuevas Manuel Alejandro <
mcs...@entel.cl> napsal:

> Taking advantage of the instance, what software do you recommend to do
> reverse engineering of postgresql?
>

please, open new thread. Your query is not related to subject.

https://wiki.postgresql.org/wiki/Design_Tools

Regards

Pavel


> Thank you
>
> Manuel
>
> El 16-11-2019, a la(s) 10:36, github kran  escribió:
>
> 
>
> *ANEXO ENTEL **Este es un correo externo: **Verifique remitente, No haga
> clic en vínculos. **Si recibes un correo malicioso avísanos a **s...@entel.cl
> ** . *
> Any reply on this please ?.
>
> On Fri, Nov 15, 2019 at 9:10 AM github kran  wrote:
>
>>
>>
>> On Thu, Nov 14, 2019 at 11:42 PM Pavel Stehule 
>> wrote:
>>
>>>   these numbers looks crazy high - how much memory has your server -
>>> more than 1TB?
>>>
>>
>> The cluster got 244 GB of RAM and storage capacity it has is 64 TB.
>>
>>>
>>>
>>> pá 15. 11. 2019 v 6:26 odesílatel github kran 
>>> napsal:
>>>

 Hello postGreSQL Community ,
>
>
>
> Hope everyone is doing great !!.
>
>
> *Background*
>
> We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for
> our day to day activities to write and read data. We have 2 clusters
> running PostgreSQL engine , one cluster
>
> keeps data up to 60 days and another cluster retains data beyond 1
> year. The data is partitioned close to a week( ~evry 5 days a partition)
> and we have around 5 partitions per month per each table and we have 2
> tables primarily so that will be 10 tables a week. So in the cluster-1 we
> have around  20 partitions and in cluster-2 we have around 160 partitions 
> (
> data from 2018). We also want to keep the data for up to 2 years in the
> cluster-2 to serve the data needs of the customer and so far we reached
> upto 1 year of maintaining this data.
>
>
>
> *Current activity*
>
> We have a custom weekly migration DB script job that moves data from 1
> cluster to another cluster what it does is the below things.
>
> 1) COPY command to copy the data from cluster-1 and split that data
> into binary files
>
> 2) Writing the binary data into the cluster-2 table
>
> 3) Creating indexes after the data is copied.
>
>
>
> *Problem what we have right now. *
>
> When the migration activity runs(weekly) from past 2 times , we saw
> the cluster read replica instance has restarted as it fallen behind the
> master(writer instance). Everything
>
> after that worked seamlessly but we want to avoid the replica getting
> restarted. To avoid from restart we started doing smaller binary files and
> copy those files to the cluster-2
>
> instead of writing 1 big file of 450 million records. We were
> successful in the recent migration as the reader instance didn’t restart
> after we split 1 big file into multiple files to copy the data over but 
> did
> restart after the indexes are created on the new table as it could be 
> write
> intensive.
>
>
>
> *DB parameters set on migration job*
>
> work_mem set to 8 GB  and maintenace_work_mem=32 GB.
>

>>>
>>>
>>> Indexes per table = 3
>
> total indexes for 2 tables = 5
>
>
>
> *DB size*
>
> Cluster-2 = 8.6 TB
>
> Cluster-1 = 3.6 TB
>
> Peak Table relational rows = 400 - 480 million rows
>
> Average table relational rows = 300 - 350 million rows.
>
> Per table size = 90 -95 GB , per table index size is about 45 GB
>
>
>
> *Questions*
>
> 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow
> down the writes to the cluster , with that the reader instance can sync 
> the
> data slowly ?.
>
> 2) Based on the above use case what are your recommendations to keep
> the data longer up to 2 years ?
>
> 3) What other recommendations you recommend ?.
>
>
>
>
>
> Appreciate your replies.
>
> THanks
> githubkran
>
>>
> --
>
> Este correo es dirigido solo a las personas que aparecen como
> destinatarios y
> puede contener información confidencial o privilegiada. Si usted recibió
> este correo
> por error, por favor notificar inmediatamente al emisor y elimine el
> correo original.
> Cualquier otro uso de este correo esta prohibido.
>
>
> *
>
> This message is for the designated recipient only and may contain
> privileged or
> confidential information. If you have received it in error, please notify
> the sender
> immediately and delete the original. Any other use of the email is
> prohibited.
>


Re: PostGreSQL Replication and question on maintenance

2019-11-20 Thread Soto Cuevas Manuel Alejandro
Taking advantage of the instance, what software do you recommend to do reverse 
engineering of postgresql?

Thank you

Manuel

El 16-11-2019, a la(s) 10:36, github kran  escribió:


ANEXO ENTEL Este es un correo externo: Verifique remitente, No haga clic en 
vínculos. Si recibes un correo malicioso avísanos a 
s...@entel.cl .
Any reply on this please ?.

On Fri, Nov 15, 2019 at 9:10 AM github kran 
mailto:githubk...@gmail.com>> wrote:


On Thu, Nov 14, 2019 at 11:42 PM Pavel Stehule 
mailto:pavel.steh...@gmail.com>> wrote:
  these numbers looks crazy high - how much memory has your server - more than 
1TB?

The cluster got 244 GB of RAM and storage capacity it has is 64 TB.


pá 15. 11. 2019 v 6:26 odesílatel github kran 
mailto:githubk...@gmail.com>> napsal:

Hello postGreSQL Community ,

Hope everyone is doing great !!.

Background
We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for our day 
to day activities to write and read data. We have 2 clusters running PostgreSQL 
engine , one cluster
keeps data up to 60 days and another cluster retains data beyond 1 year. The 
data is partitioned close to a week( ~evry 5 days a partition) and we have 
around 5 partitions per month per each table and we have 2 tables primarily so 
that will be 10 tables a week. So in the cluster-1 we have around  20 
partitions and in cluster-2 we have around 160 partitions ( data from 2018). We 
also want to keep the data for up to 2 years in the cluster-2 to serve the data 
needs of the customer and so far we reached upto 1 year of maintaining this 
data.

Current activity
We have a custom weekly migration DB script job that moves data from 1 cluster 
to another cluster what it does is the below things.
1) COPY command to copy the data from cluster-1 and split that data into binary 
files
2) Writing the binary data into the cluster-2 table
3) Creating indexes after the data is copied.

Problem what we have right now.
When the migration activity runs(weekly) from past 2 times , we saw the cluster 
read replica instance has restarted as it fallen behind the master(writer 
instance). Everything
after that worked seamlessly but we want to avoid the replica getting 
restarted. To avoid from restart we started doing smaller binary files and copy 
those files to the cluster-2
instead of writing 1 big file of 450 million records. We were successful in the 
recent migration as the reader instance didn’t restart after we split 1 big 
file into multiple files to copy the data over but did restart after the 
indexes are created on the new table as it could be write intensive.

DB parameters set on migration job
work_mem set to 8 GB  and maintenace_work_mem=32 GB.



Indexes per table = 3
total indexes for 2 tables = 5

DB size
Cluster-2 = 8.6 TB
Cluster-1 = 3.6 TB
Peak Table relational rows = 400 - 480 million rows
Average table relational rows = 300 - 350 million rows.
Per table size = 90 -95 GB , per table index size is about 45 GB

Questions
1) Can we decrease the maintenace_work_mem to 16 GB and will it slow down the 
writes to the cluster , with that the reader instance can sync the data slowly 
?.
2) Based on the above use case what are your recommendations to keep the data 
longer up to 2 years ?
3) What other recommendations you recommend ?.


Appreciate your replies.

THanks
githubkran



Este correo es dirigido solo a las personas que aparecen como destinatarios y
puede contener información confidencial o privilegiada. Si usted recibió este 
correo
por error, por favor notificar inmediatamente al emisor y elimine el correo 
original.
Cualquier otro uso de este correo esta prohibido.

*

This message is for the designated recipient only and may contain privileged or
confidential information. If you have received it in error, please notify the 
sender
immediately and delete the original. Any other use of the email is prohibited.


Re: PostGreSQL Replication and question on maintenance

2019-11-16 Thread Jeff Janes
On Thu, Nov 14, 2019 at 12:23 PM github kran  wrote:

>
>>
>> *Problem what we have right now. *
>>
>> When the migration activity runs(weekly) from past 2 times , we saw the
>> cluster read replica instance has restarted as it fallen behind the
>> master(writer instance).
>>
>
I can't figure out what your setup is here.  You must be using logical
replication (not physical) or you wouldn't be able to write to the replica
at all.  But if you are using logical replication, why do you also need
these weekly jobs?  Why isn't logical replication taking care of it?




> Everything
>>
>> after that worked seamlessly but we want to avoid the replica getting
>> restarted. To avoid from restart we started doing smaller binary files and
>> copy those files to the cluster-2
>>
>
Who restarted it?  I am not aware of any case where the replica responds to
falling behind by restarting itself.  With physical replication, it can
start cancelling queries, but you don't seem to be using physical
replication.

Cheers,

Jeff

>


Re: PostGreSQL Replication and question on maintenance

2019-11-16 Thread github kran
Any reply on this please ?.

On Fri, Nov 15, 2019 at 9:10 AM github kran  wrote:

>
>
> On Thu, Nov 14, 2019 at 11:42 PM Pavel Stehule 
> wrote:
>
>>   these numbers looks crazy high - how much memory has your server - more
>> than 1TB?
>>
>
> The cluster got 244 GB of RAM and storage capacity it has is 64 TB.
>
>>
>>
>> pá 15. 11. 2019 v 6:26 odesílatel github kran 
>> napsal:
>>
>>>
>>> Hello postGreSQL Community ,



 Hope everyone is doing great !!.


 *Background*

 We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for
 our day to day activities to write and read data. We have 2 clusters
 running PostgreSQL engine , one cluster

 keeps data up to 60 days and another cluster retains data beyond 1
 year. The data is partitioned close to a week( ~evry 5 days a partition)
 and we have around 5 partitions per month per each table and we have 2
 tables primarily so that will be 10 tables a week. So in the cluster-1 we
 have around  20 partitions and in cluster-2 we have around 160 partitions (
 data from 2018). We also want to keep the data for up to 2 years in the
 cluster-2 to serve the data needs of the customer and so far we reached
 upto 1 year of maintaining this data.



 *Current activity*

 We have a custom weekly migration DB script job that moves data from 1
 cluster to another cluster what it does is the below things.

 1) COPY command to copy the data from cluster-1 and split that data
 into binary files

 2) Writing the binary data into the cluster-2 table

 3) Creating indexes after the data is copied.



 *Problem what we have right now. *

 When the migration activity runs(weekly) from past 2 times , we saw the
 cluster read replica instance has restarted as it fallen behind the
 master(writer instance). Everything

 after that worked seamlessly but we want to avoid the replica getting
 restarted. To avoid from restart we started doing smaller binary files and
 copy those files to the cluster-2

 instead of writing 1 big file of 450 million records. We were
 successful in the recent migration as the reader instance didn’t restart
 after we split 1 big file into multiple files to copy the data over but did
 restart after the indexes are created on the new table as it could be write
 intensive.



 *DB parameters set on migration job*

 work_mem set to 8 GB  and maintenace_work_mem=32 GB.

>>>
>>
>>
>> Indexes per table = 3

 total indexes for 2 tables = 5



 *DB size*

 Cluster-2 = 8.6 TB

 Cluster-1 = 3.6 TB

 Peak Table relational rows = 400 - 480 million rows

 Average table relational rows = 300 - 350 million rows.

 Per table size = 90 -95 GB , per table index size is about 45 GB



 *Questions*

 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow
 down the writes to the cluster , with that the reader instance can sync the
 data slowly ?.

 2) Based on the above use case what are your recommendations to keep
 the data longer up to 2 years ?

 3) What other recommendations you recommend ?.





 Appreciate your replies.

 THanks
 githubkran

>


Re: PostGreSQL Replication and question on maintenance

2019-11-15 Thread github kran
On Thu, Nov 14, 2019 at 11:42 PM Pavel Stehule 
wrote:

>   these numbers looks crazy high - how much memory has your server - more
> than 1TB?
>

The cluster got 244 GB of RAM and storage capacity it has is 64 TB.

>
>
> pá 15. 11. 2019 v 6:26 odesílatel github kran 
> napsal:
>
>>
>> Hello postGreSQL Community ,
>>>
>>>
>>>
>>> Hope everyone is doing great !!.
>>>
>>>
>>> *Background*
>>>
>>> We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for
>>> our day to day activities to write and read data. We have 2 clusters
>>> running PostgreSQL engine , one cluster
>>>
>>> keeps data up to 60 days and another cluster retains data beyond 1 year.
>>> The data is partitioned close to a week( ~evry 5 days a partition) and we
>>> have around 5 partitions per month per each table and we have 2 tables
>>> primarily so that will be 10 tables a week. So in the cluster-1 we have
>>> around  20 partitions and in cluster-2 we have around 160 partitions ( data
>>> from 2018). We also want to keep the data for up to 2 years in the
>>> cluster-2 to serve the data needs of the customer and so far we reached
>>> upto 1 year of maintaining this data.
>>>
>>>
>>>
>>> *Current activity*
>>>
>>> We have a custom weekly migration DB script job that moves data from 1
>>> cluster to another cluster what it does is the below things.
>>>
>>> 1) COPY command to copy the data from cluster-1 and split that data into
>>> binary files
>>>
>>> 2) Writing the binary data into the cluster-2 table
>>>
>>> 3) Creating indexes after the data is copied.
>>>
>>>
>>>
>>> *Problem what we have right now. *
>>>
>>> When the migration activity runs(weekly) from past 2 times , we saw the
>>> cluster read replica instance has restarted as it fallen behind the
>>> master(writer instance). Everything
>>>
>>> after that worked seamlessly but we want to avoid the replica getting
>>> restarted. To avoid from restart we started doing smaller binary files and
>>> copy those files to the cluster-2
>>>
>>> instead of writing 1 big file of 450 million records. We were successful
>>> in the recent migration as the reader instance didn’t restart after we
>>> split 1 big file into multiple files to copy the data over but did restart
>>> after the indexes are created on the new table as it could be write
>>> intensive.
>>>
>>>
>>>
>>> *DB parameters set on migration job*
>>>
>>> work_mem set to 8 GB  and maintenace_work_mem=32 GB.
>>>
>>
>
>
> Indexes per table = 3
>>>
>>> total indexes for 2 tables = 5
>>>
>>>
>>>
>>> *DB size*
>>>
>>> Cluster-2 = 8.6 TB
>>>
>>> Cluster-1 = 3.6 TB
>>>
>>> Peak Table relational rows = 400 - 480 million rows
>>>
>>> Average table relational rows = 300 - 350 million rows.
>>>
>>> Per table size = 90 -95 GB , per table index size is about 45 GB
>>>
>>>
>>>
>>> *Questions*
>>>
>>> 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow
>>> down the writes to the cluster , with that the reader instance can sync the
>>> data slowly ?.
>>>
>>> 2) Based on the above use case what are your recommendations to keep the
>>> data longer up to 2 years ?
>>>
>>> 3) What other recommendations you recommend ?.
>>>
>>>
>>>
>>>
>>>
>>> Appreciate your replies.
>>>
>>> THanks
>>> githubkran
>>>



Re: PostGreSQL Replication and question on maintenance

2019-11-14 Thread Pavel Stehule
pá 15. 11. 2019 v 6:26 odesílatel github kran  napsal:

>
> Hello postGreSQL Community ,
>>
>>
>>
>> Hope everyone is doing great !!.
>>
>>
>> *Background*
>>
>> We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for our
>> day to day activities to write and read data. We have 2 clusters running
>> PostgreSQL engine , one cluster
>>
>> keeps data up to 60 days and another cluster retains data beyond 1 year.
>> The data is partitioned close to a week( ~evry 5 days a partition) and we
>> have around 5 partitions per month per each table and we have 2 tables
>> primarily so that will be 10 tables a week. So in the cluster-1 we have
>> around  20 partitions and in cluster-2 we have around 160 partitions ( data
>> from 2018). We also want to keep the data for up to 2 years in the
>> cluster-2 to serve the data needs of the customer and so far we reached
>> upto 1 year of maintaining this data.
>>
>>
>>
>> *Current activity*
>>
>> We have a custom weekly migration DB script job that moves data from 1
>> cluster to another cluster what it does is the below things.
>>
>> 1) COPY command to copy the data from cluster-1 and split that data into
>> binary files
>>
>> 2) Writing the binary data into the cluster-2 table
>>
>> 3) Creating indexes after the data is copied.
>>
>>
>>
>> *Problem what we have right now. *
>>
>> When the migration activity runs(weekly) from past 2 times , we saw the
>> cluster read replica instance has restarted as it fallen behind the
>> master(writer instance). Everything
>>
>> after that worked seamlessly but we want to avoid the replica getting
>> restarted. To avoid from restart we started doing smaller binary files and
>> copy those files to the cluster-2
>>
>> instead of writing 1 big file of 450 million records. We were successful
>> in the recent migration as the reader instance didn’t restart after we
>> split 1 big file into multiple files to copy the data over but did restart
>> after the indexes are created on the new table as it could be write
>> intensive.
>>
>>
>>
>> *DB parameters set on migration job*
>>
>> work_mem set to 8 GB  and maintenace_work_mem=32 GB.
>>
>
these numbers looks crazy high - how much memory has your server - more
than 1TB?


Indexes per table = 3
>>
>> total indexes for 2 tables = 5
>>
>>
>>
>> *DB size*
>>
>> Cluster-2 = 8.6 TB
>>
>> Cluster-1 = 3.6 TB
>>
>> Peak Table relational rows = 400 - 480 million rows
>>
>> Average table relational rows = 300 - 350 million rows.
>>
>> Per table size = 90 -95 GB , per table index size is about 45 GB
>>
>>
>>
>> *Questions*
>>
>> 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow down
>> the writes to the cluster , with that the reader instance can sync the data
>> slowly ?.
>>
>> 2) Based on the above use case what are your recommendations to keep the
>> data longer up to 2 years ?
>>
>> 3) What other recommendations you recommend ?.
>>
>>
>>
>>
>>
>> Appreciate your replies.
>>
>> THanks
>> githubkran
>>
>>>


Fwd: PostGreSQL Replication and question on maintenance

2019-11-14 Thread github kran
> Hello postGreSQL Community ,
>
>
>
> Hope everyone is doing great !!.
>
>
> *Background*
>
> We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for our
> day to day activities to write and read data. We have 2 clusters running
> PostgreSQL engine , one cluster
>
> keeps data up to 60 days and another cluster retains data beyond 1 year.
> The data is partitioned close to a week( ~evry 5 days a partition) and we
> have around 5 partitions per month per each table and we have 2 tables
> primarily so that will be 10 tables a week. So in the cluster-1 we have
> around  20 partitions and in cluster-2 we have around 160 partitions ( data
> from 2018). We also want to keep the data for up to 2 years in the
> cluster-2 to serve the data needs of the customer and so far we reached
> upto 1 year of maintaining this data.
>
>
>
> *Current activity*
>
> We have a custom weekly migration DB script job that moves data from 1
> cluster to another cluster what it does is the below things.
>
> 1) COPY command to copy the data from cluster-1 and split that data into
> binary files
>
> 2) Writing the binary data into the cluster-2 table
>
> 3) Creating indexes after the data is copied.
>
>
>
> *Problem what we have right now. *
>
> When the migration activity runs(weekly) from past 2 times , we saw the
> cluster read replica instance has restarted as it fallen behind the
> master(writer instance). Everything
>
> after that worked seamlessly but we want to avoid the replica getting
> restarted. To avoid from restart we started doing smaller binary files and
> copy those files to the cluster-2
>
> instead of writing 1 big file of 450 million records. We were successful
> in the recent migration as the reader instance didn’t restart after we
> split 1 big file into multiple files to copy the data over but did restart
> after the indexes are created on the new table as it could be write
> intensive.
>
>
>
> *DB parameters set on migration job*
>
> work_mem set to 8 GB  and maintenace_work_mem=32 GB.
>
> Indexes per table = 3
>
> total indexes for 2 tables = 5
>
>
>
> *DB size*
>
> Cluster-2 = 8.6 TB
>
> Cluster-1 = 3.6 TB
>
> Peak Table relational rows = 400 - 480 million rows
>
> Average table relational rows = 300 - 350 million rows.
>
> Per table size = 90 -95 GB , per table index size is about 45 GB
>
>
>
> *Questions*
>
> 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow down
> the writes to the cluster , with that the reader instance can sync the data
> slowly ?.
>
> 2) Based on the above use case what are your recommendations to keep the
> data longer up to 2 years ?
>
> 3) What other recommendations you recommend ?.
>
>
>
>
>
> Appreciate your replies.
>
> THanks
> githubkran
>
>>


PostGreSQL Replication and question on maintenance

2019-11-14 Thread github kran
sorry changing the subject line.

On Thu, Nov 14, 2019 at 11:21 AM github kran  wrote:

> Hello Team,
>
>
>
> Hope everyone is doing great !!.
>
>
> *Background*
>
> We use PostgreSQL Version 10.6 version and heavily use PostgreSQL for our
> day to day activities to write and read data. We have 2 clusters running
> PostgreSQL engine , one cluster
>
> keeps data up to 60 days and another cluster retains data beyond 1 year.
> The data is partitioned close to a week( ~evry 5 days a partition) and we
> have around 5 partitions per month per each table and we have 2 tables
> primarily so that will be 10 tables a week. So in the cluster-1 we have
> around  20 partitions and in cluster-2 we have around 160 partitions ( data
> from 2018). We also want to keep the data for up to 2 years in the
> cluster-2 to serve the data needs of the customer and so far we reached
> upto 1 year of maintaining this data.
>
>
>
> *Current activity*
>
> We have a custom weekly migration DB script job that moves data from 1
> cluster to another cluster what it does is the below things.
>
> 1) COPY command to copy the data from cluster-1 and split that data into
> binary files
>
> 2) Writing the binary data into the cluster-2 table
>
> 3) Creating indexes after the data is copied.
>
>
>
> *Problem what we have right now. *
>
> When the migration activity runs(weekly) from past 2 times , we saw the
> cluster read replica instance has restarted as it fallen behind the
> master(writer instance). Everything
>
> after that worked seamlessly but we want to avoid the replica getting
> restarted. To avoid from restart we started doing smaller binary files and
> copy those files to the cluster-2
>
> instead of writing 1 big file of 450 million records. We were successful
> in the recent migration as the reader instance didn’t restart after we
> split 1 big file into multiple files to copy the data over but did restart
> after the indexes are created on the new table as it could be write
> intensive.
>
>
>
> *DB parameters set on migration job*
>
> work_mem set to 8 GB  and maintenace_work_mem=32 GB.
>
> Indexes per table = 3
>
> total indexes for 2 tables = 5
>
>
>
> *DB size*
>
> Cluster-2 = 8.6 TB
>
> Cluster-1 = 3.6 TB
>
> Peak Table relational rows = 400 - 480 million rows
>
> Average table relational rows = 300 - 350 million rows.
>
> Per table size = 90 -95 GB , per table index size is about 45 GB
>
>
>
> *Questions*
>
> 1) Can we decrease the maintenace_work_mem to 16 GB and will it slow down
> the writes to the cluster , with that the reader instance can sync the data
> slowly ?.
>
> 2) Based on the above use case what are your recommendations to keep the
> data longer up to 2 years ?
>
> 3) What other recommendations you recommend ?.
>
>
>
>
>
> Appreciate your replies.
>
> THanks
> githubkran
>
>>