Prepare for the PgBouncer and IPv4 deprecations on 26th January 2024


Read Replicas

Spin up additional read-only databases for your project

Read replicas continuously replicate data from a primary database. You can both read and write data on a primary database, but you can only read (for example, use select statements), and not write, on a read replica.

Read Replica---

When you create a project, the automatically created database is considered the primary database. If you then add read replicas, they will receive and synchronize data from this primary database. Replication is asynchronous. It happens in the background, so transactions on the primary aren't blocked. The time delay between writing data to the primary and the read replica receiving the change is called replication lag.

Read replicas have many applications:

  • They reduce read load on the primary database.
    • For example, you can use a read replica for complex analytical queries and reserve the primary for user-facing create, update, and delete operations.
  • For projects that need to cover a global audience, additional databases can be spun up closer to users in any of our twelve supported regions to reduce latency.
  • Read replicas provide data redundancy.

How are Read Replicas made?

We use a hybrid approach to replicate data from a primary to its read replicas, combining the native methods of streaming replication and file-based log shipping.

Streaming replication

Postgres generates a Write Ahead Log (WAL) as database changes occur. With streaming replication, these changes stream from the primary to the read replica server. The WAL alone is sufficient to reconstruct the database to its current state.

This replication method is fast, since changes are streamed directly from the primary to the read replica. On the other hand, it faces challenges when the read replica can't keep up with the WAL changes from its primary. This can happen when the read replica is too small, running on degraded hardware, or has a heavier workload running.

To address this, Postgres does provide tunable configuration, like wal_keep_size, to adjust the WAL retained by the primary. If the read replica fails to “catch up” before the WAL surpasses the wal_keep_size setting, the replication is terminated. Tuning is a bit of an art - the amount of WAL required is variable for every situation.

File-based log shipping

In this replication method, the primary continuously buffers WAL changes to a local file and then sends the file to the read replica. If multiple read replicas are present, files could also be sent to an intermediary location accessible by all. The read replica then reads the WAL files and applies those changes. There is higher replication lag than streaming replication since the primary buffers the changes locally first. It also means there is a small chance that WAL changes do not reach read replicas if the primary goes down before the file is transferred. In these cases, if the primary fails a replica using streaming replication would (in most cases) be more up-to-date than a replica using file-based log shipping.

File-based log shipping 🤝 streaming replication

How replication is achieved.

We bring these two methods together to achieve quick, stable, and reliable replication. Each method addresses the limitations of the other. Streaming replication minimizes replication lag, while file-based log shipping provides a fallback. For file-based log shipping, we use our existing Point In Time Recovery (PITR) infrastructure. We regularly archive files from the primary using WAL-G, an open source archival and restoration tool, and ship the WAL files to S3.

We combine it with streaming replication to reduce replication lag. Once WAL-G files have been synced from S3, read replicas connect to the primary and stream the WAL directly.

Getting started

Read replicas can be managed from the infrastructure settings page. For now, each project can only spin up to two read replicas. Read replicas also inherit the compute size of their primary database.

Read replicas offer the following features:

Dedicated endpoints

Each read replica has its own dedicated database and API endpoints. The database endpoint can be found in your dashboard database settings under Connection info. The API endpoint can be found in dashboard API settings under Project URL.

A read replica's API endpoint only supports GET requests to the REST API. Requests towards other Supabase products, such as Auth, Storage, and Realtime are not able to utilize a read replica nor its API endpoint for now. We are working on adding read replica support for them.

Dedicated connection pool

A connection pool through Supavisor is also available for each read replica. Its unique connection string can be found in your dashboard database settings under Connection string.

Querying through the SQL editor

In the SQL editor, you can choose if you want to run the query on the primary or one of the replicas.

API load balancer

When read replicas are spun up, an API load balancer becomes available for the project. An endpoint for it is provisioned and can be found here.

The load balancer uses a round-robin strategy to route GET requests to any of the API endpoints available, including that of the primary database. Non-GET requests can also be sent through this endpoint. They are immediately redirected to the primary database.

Similar to the read replica API endpoints, only requests to the REST API are available for now. Support for other Supabase products will become available once read replica support has been completed for them.

When a project no longer has any read replicas, the load balancer and its endpoint are brought down as well.

Centralized configuration management

All settings made through the dashboard will be propagated across all databases of a project. This ensures that no read replica get out of sync with its primary database or with other read replicas.


Prerequisites for using read replicas

A project needs to satisfy these requirements to spin up a read replica:

  1. Running on AWS.
    • Support for projects on will be added at a later date.
  2. Running on at least a Small compute add-on.
  3. Using physical backups
    • Physical backups are automatically enabled if using PITR
  4. Running on Postgres 15.

Read replicas need to be brought down before starting the following procedures

The following procedures require all read replicas for a project to be brought down before they can be started:

  1. Upgrading a project
  2. Performing data restoration.

Once these operations complete, read replicas can be spun back up.

Compute Add-ons

In order to avoid replicas being unable to keep up with activity on the primaries they're following, Read Replicas are currently placed on the same Compute Add-on as the primary they're following.