Known issues v4
This section discusses currently known issues in EDB Postgres Distributed 4. These issues are tracked in BDR's ticketing system and are expected to be resolved in a future release.
Performance of HARP in terms of failover and switchover time depends non-linearly on the latencies between DCS nodes. Which is why we currently recommend using etcd cluster per region for HARP in case of EDB Postgres Distributed deployment over multiple regions (typically the Gold and Platinum layouts). TPAexec already sets up the etcd do run per region cluster for these when
harp_consensus_protocol
option is set toetcd
in theconfig.yml
. It's recommended to increase theleader_lease_duration
HARP option (harp_leader_lease_duration
in TPAexec) for DCS deployments across higher latency network.If the resolver for the
update_origin_change
conflict is set toskip
,synchronous_commit=remote_apply
is used, and concurrent updates of the same row are repeatedly applied on two different nodes, then one of the update statements might hang due to a deadlock with the BDR writer. As mentioned in the Conflicts chapter,skip
is not the default resolver for theupdate_origin_change
conflict, and this combination isn't intended to be used in production. It discards one of the two conflicting updates based on the order of arrival on that node, which is likely to cause a divergent cluster. In the rare situation that you do choose to use theskip
conflict resolver, note the issue with the use of theremote_apply
mode.The Decoding Worker feature doesn't work with CAMO/EAGER/Group Commit. Installations using CAMO/Eager/Group Commit must keep
enable_wal_decoder
disabled.Decoding Worker works only with the default replication sets.
Lag control doesn't adjust commit delay in any way on a fully isolated node, that is, in case all other nodes are unreachable or not operational. As soon as at least one node is connected, replication lag control picks up its work and adjusts the BDR commit delay again.
For time-based lag control, BDR currently uses the lag time (measured by commit timestamps) rather than the estimated catchup time that's based on historic apply rate.
Changing the CAMO partners in a CAMO pair isn't currently possible. It's possible only to add or remove a pair. Adding or removing a pair doesn't need a restart of Postgres or even a reload of the configuration.
Group Commit cannot be combined with CAMO or Eager All Node replication. Eager Replication currently only works by using the "global" BDR commit scope.
Neither Eager replication nor Group Commit support
synchronous_replication_availability = 'async'
.Group Commit doesn't support a timeout of the commit after
bdr.global_commit_timeout
.Transactions using Eager Replication can't yet execute DDL, nor do they support explicit two-phase commit. The TRUNCATE command is allowed.
Not all DDL can be run when either CAMO or Group Commit is used.
Parallel apply is not currently supported in combination with Group Commit, please make sure to disable it when using Group Commit by either setting
num_writers
to 1 for the node group (usingbdr.alter_node_group_config
) or via the GUCbdr.writers_per_subscription
(see Configuration of Generic Replication).There currently is no protection against altering or removing a commit scope. Running transactions in a commit scope that is concurrently being altered or removed can lead to the transaction blocking or replication stalling completely due to an error on the downstream node attempting to apply the transaction. Ensure that any transactions using a specific commit scope have finished before altering or removing it.
Details of other design or implementation limitations are also available.