Online PostgreSQL Database Migration
In this exercise you will configure logical replication between a source PostgreSQL Server and Azure Database for PostgreSQL Flexible Server to allow for an online migration activity to take place.
Before you start
You need your own Azure subscription to complete this exercise. If you don’t have an Azure subscription, you can create an Azure free trial.
Note:: This exercise will require that the server you use as a source for the migration is accessible to the Azure Database for PostgreSQL Flexible Server so that it can connect and migrate databases. This will require that the source server is accessible via a public IP address and port. > A list of Azure Region IP Addresses can be downloaded from Azure IP Ranges and Service Tags – Public Cloud to help minimize the allowed ranges of IP Addresses in your firewall rules based on the Azure region used.
Open your server’s firewall to allow the Migration feature within the Azure Database for PostgreSQL Flexible Server access to the source PostgreSQL Server, which by default is TCP port 5432.
When using a firewall appliance in front of your source database, you may need to add firewall rules to allow the Migration feature within the Azure Database for PostgreSQL Flexible Server to access the source database(s) for migration.
The maximum supported version of PostgreSQL for migration is version 16.
Prerequisites
Note: Before starting this exercise you will need to have completed the previous exercise to have the source and target databases in place to configure logical replication as this exercise builds on the activity in that one.
Create Publication - Source Server
- Open PGAdmin and connect to the source server which contains the database which is going to act as the source for the data synchronization to the Azure Database for PostgreSQL Flexible Server.
- Open a new Query window connected to the source database with the data we want to synchronize.
- Configure the source server wal_level to logical to allow for publication of data.
- Locate and open the postgresql.conf file in the bin directory within the PostgreSQL installation directory.
- Find the line with the configuration setting wal_level.
- Ensure that the line is un-commented and set the value to logical.
- Save and close the file.
- Restart the PostgreSQL Service.
-
Now configure a publication which will contain all of the tables within the database.
CREATE PUBLICATION migration1 FOR ALL TABLES;
Create Subscription - Target Server
- Open PGAdmin and connect to the Azure Database for PostgreSQL Flexible Server which contains the database which is going to act as the target for the data synchronization from the source server.
- Open a new Query window connected to the source database with the data we want to synchronize.
-
Create the subscription to the source server.
CREATE SUBSCRIPTION migration1 CONNECTION 'host=<source server name> port=<server port> dbname=adventureworks application_name=migration1 user=<username> password=<password>' PUBLICATION migration1 WITH(copy_data = false) ;
-
Check the status of the table replication.
SELECT PT.schemaname, PT.tablename, CASE PS.srsubstate WHEN 'i' THEN 'initialize' WHEN 'd' THEN 'data is being copied' WHEN 'f' THEN 'finished table copy' WHEN 's' THEN 'synchronized' WHEN 'r' THEN ' ready (normal replication)' ELSE 'unknown' END AS replicationState FROM pg_publication_tables PT, pg_subscription_rel PS JOIN pg_class C ON (C.oid = PS.srrelid) JOIN pg_namespace N ON (N.oid = C.relnamespace) ;
Test Data Replication
-
On the Source Server check the row count of the workorder table.
SELECT COUNT(*) FROM production.workorder;
-
On the Target Server check the row count of the workorder table.
SELECT COUNT(*) FROM production.workorder;
- Check that the row count values match.
- Now download the Lab11_workorder.csv file from the repository here to C:\
-
Load new data into the workorder table on the source server from the CSV using the following command.
psql --host=localhost --port=5432 --username=postgres --dbname=adventureworks --command="\COPY production.workorder FROM 'C:\Lab11_workorder.csv' CSV HEADER"
The command output should be COPY 490
, indicating that 490 additional rows were written into the table from the CSV file.
- Check the row counts for the workorder table in the source (72591 rows) and destination match to verify that the data replication is working.
Exercise Clean-up
The Azure Database for PostgreSQL we deployed in this exercise will incur charges you can delete the server after this exercise. Alternatively, you can delete the rg-learn-work-with-postgresql-eastus resource group to remove all resources that we deployed as part of this exercise.