Postgres replication
Postgres is one of the most popular relational database management systems. This verified source uses Postgres replication functionality to efficiently process tables (a process often referred to as Change Data Capture or CDC). It uses logical decoding and the standard built-in pgoutput
output plugin.
Resources that can be loaded using this verified source are:
Name | Description |
---|---|
replication_resource | Load published messages from a replication slot |
The postgres replication source currently does not suppport the scd2 merge strategy.
Setup Guide
Setup user
To setup a Postgres user follow these steps:
The Postgres user needs to have the
LOGIN
andREPLICATION
attributes assigned:CREATE ROLE replication_user WITH LOGIN REPLICATION;
It also needs
GRANT
privilege on the database:GRANT CREATE ON DATABASE dlt_data TO replication_user;
Set up RDS
To setup a Postgres user on RDS follow these steps:
You must enable replication for RDS Postgres instance via Parameter Group
WITH LOGIN REPLICATION;
does not work on RDS, instead do:GRANT rds_replication TO replication_user;
Do not fallback to non-SSL connection by setting connection parameters:
sources.pg_replication.credentials="postgresql://loader:password@host.rds.amazonaws.com:5432/dlt_data?sslmode=require&connect_timeout=300"
Initialize the verified source
To get started with your data pipeline, follow these steps:
Enter the following command:
dlt init pg_replication duckdb
It will initialize the pipeline example with a Postgres replication as the source and DuckDB as the destination.
If you'd like to use a different destination, simply replace
duckdb
with the name of your preferred destination.This source uses
sql_database
source, you can init it as follows:dlt init sql_database duckdb
noteIt is important to note that It is now only required if a user performs an initial load, specifically when
persist_snapshots
is set toTrue
.After running these two commands, a new directory will be created with the necessary files and configuration settings to get started.
For more information, read the guide on how to add a verified source.
noteYou can omit the
[sql.sources.credentials]
section insecrets.toml
as it is not required.
Add credentials
In the
.dlt
folder, there's a file calledsecrets.toml
. It's where you store sensitive information securely, like access tokens. Keep this file safe.Here's what the
secrets.toml
looks like:[sources.pg_replication.credentials]
drivername = "postgresql" # please set me up!
database = "database" # please set me up!
password = "password" # please set me up!
username = "username" # please set me up!
host = "host" # please set me up!
port = 0 # please set me up!Credentials can be set as shown above. Alternatively, you can provide credentials in the
secrets.toml
file as follows:sources.pg_replication.credentials="postgresql://username@password.host:port/database"
Finally, follow the instructions in Destinations to add credentials for your chosen destination. This will ensure that your data is properly routed.
For more information, read the Configuration section.
Run the pipeline
- Before running the pipeline, ensure that you have installed all the necessary dependencies by running the command:
pip install -r requirements.txt
- You're now ready to run the pipeline! To get started, run the following command:
python pg_replication_pipeline.py
- Once the pipeline has finished running, you can verify that everything loaded correctly by using the following command:For example, the
dlt pipeline <pipeline_name> show
pipeline_name
for the above pipeline example ispg_replication_pipeline
, you may also use any custom name instead.
For more information, read the guide on how to run a pipeline.
Sources and resources
dlt
works on the principle of sources and resources.
Resource replication_resource
This resource yields data items for changes in one or more Postgres tables.
@dlt.resource(
name=lambda args: args["slot_name"] + "_" + args["pub_name"],
standalone=True,
)
def replication_resource(
slot_name: str,
pub_name: str,
credentials: ConnectionStringCredentials = dlt.secrets.value,
include_columns: Optional[Dict[str, Sequence[str]]] = None,
columns: Optional[Dict[str, TTableSchemaColumns]] = None,
target_batch_size: int = 1000,
flush_slot: bool = True,
) -> Iterable[Union[TDataItem, DataItemWithMeta]]:
...
slot_name
: Replication slot name to consume messages.
pub_name
: Publication slot name to publish messages.
include_columns
: Maps table name(s) to sequence of names of columns to include in the generated data items. Any column not in the sequence is excluded. If not provided, all columns are included
columns
: Maps table name(s) to column hints to apply on the replicated table(s)
target_batch_size
: Desired number of data items yielded in a batch. Can be used to limit the data items in memory.
flush_slot
: Whether processed messages are discarded from the replication slot. The recommended value is "True".
Customization
If you wish to create your own pipelines, you can leverage source and resource methods from this verified source.
Define the source pipeline as:
# Defining source pipeline
src_pl = dlt.pipeline(
pipeline_name="source_pipeline",
destination="postgres",
dataset_name="source_dataset",
dev_mode=True,
)You can configure and use the
get_postgres_pipeline()
function available in thepg_replication_pipeline.py
file to achieve the same functionality.IMPORTANTWhen working with large datasets from a Postgres database, it's important to consider the relevance of the source pipeline. For testing purposes, using the source pipeline can be beneficial to try out the data flow. However, in production use cases, there will likely be another process that mutates the Postgres database. In such cases, the user generally only needs to define a destination pipeline.
Similarly, define the destination pipeline.
dest_pl = dlt.pipeline(
pipeline_name="pg_replication_pipeline",
destination='duckdb',
dataset_name="replicate_single_table",
dev_mode=True,
)Define the slot and publication names as:
slot_name = "example_slot"
pub_name = "example_pub"To initialize replication, you can use the
init_replication
function. A user can use this function to letdlt
configure Postgres and make it ready for replication.# requires the Postgres user to have the REPLICATION attribute assigned
init_replication(
slot_name=slot_name,
pub_name=pub_name,
schema_name=src_pl.dataset_name,
table_names="my_source_table",
reset=True,
)noteTo replicate the entire schema, you can omit the
table_names
argument from theinit_replication
function.To snapshot the data to the destination during the initial load, you can use the
persist_snapshots=True
argument as follows:snapshot = init_replication( # requires the Postgres user to have the REPLICATION attribute assigned
slot_name=slot_name,
pub_name=pub_name,
schema_name=src_pl.dataset_name,
table_names="my_source_table",
persist_snapshots=True, # persist snapshot table(s) and let function return resource(s) for initial load
reset=True,
)To load this snapshot to the destination, run the destination pipeline as:
dest_pl.run(snapshot)
After changes are made to the source, you can replicate the changes to the destination using the
replication_resource
, and run the pipeline as:# Create a resource that generates items for each change in the source table
changes = replication_resource(slot_name, pub_name)
# Run the pipeline as
dest_pl.run(changes)To replicate tables with selected columns you can use the
include_columns
argument as follows:# requires the Postgres user to have the REPLICATION attribute assigned
initial_load = init_replication(
slot_name=slot_name,
pub_name=pub_name,
schema_name=src_pl.dataset_name,
table_names="my_source_table",
include_columns={
"my_source_table": ("column1", "column2")
},
reset=True,
)Similarly, to replicate changes from selected columns, you can use the
table_names
andinclude_columns
arguments in thereplication_resource
function.