Hi,
Its postgres database behind the scenes.

We have a use case in which the customer is planning to migrate data from
an older version (V1) to a newer version (V2). For V2, the tables will be
new, but their structure will be similar to the V1 version with few changes
in relationship might be there. We want to have this migration approach
happen in multiple phases in which each time the delta data from version V1
will be moved to version- V2 and then final cutover will  happen to V2 if
all looks good or else rollback to V1. The tables are smaller in size like
max ~100K records in tables.

My question is, is it a good idea to have an approach in which we will have
procedures created to move the delta data in every phase and schedule those
using some tasks for each table. Or any other strategy should we follow?

Also another thing to note , we have used sequences as primary keys in some
tables and they have FK relationships with other tables, so the same
sequence number in version V2 will cause issues/conflict, so how should we
handle this scenario? Should we just create new sequences with higher start
values?

Regards
Veem

Reply via email to