This is the multi-page printable view of this section. Click here to print.

Return to the regular view of this page.

Module: ETCD

Pigsty deploys etcd as DCS for reliable distributed config storage, supporting PostgreSQL HA.

ETCD is a distributed, reliable key-value store for critical system config data.

Pigsty uses etcd as DCS (Distributed Config Store), critical for PostgreSQL HA and automatic failover.

The ETCD module depends on NODE module and is required by PGSQL module. Install NODE module to manage nodes before installing ETCD.

Deploy ETCD cluster before any PGSQL cluster—patroni and vip-manager for PG HA rely on etcd for HA and L2 VIP binding to primary.

flowchart LR
    subgraph PGSQL [PGSQL]
        patroni[Patroni]
        vip[VIP Manager]
    end

    subgraph ETCD [ETCD]
        etcd[DCS Service]
    end

    subgraph NODE [NODE]
        node[Software Repo]
    end

    PGSQL -->|depends| ETCD -->|depends| NODE

    style PGSQL fill:#3E668F,stroke:#2d4a66,color:#fff
    style ETCD fill:#5B9CD5,stroke:#4178a8,color:#fff
    style NODE fill:#FCDB72,stroke:#d4b85e,color:#333

    style patroni fill:#2d4a66,stroke:#1e3347,color:#fff
    style vip fill:#2d4a66,stroke:#1e3347,color:#fff
    style etcd fill:#4178a8,stroke:#2d5a7a,color:#fff
    style node fill:#d4b85e,stroke:#b89a4a,color:#333

One etcd cluster per Pigsty deployment serves multiple PG clusters.

Pigsty enables RBAC by default. Each PG cluster uses independent credentials for multi-tenant isolation. Admins use etcd root user with full permissions over all PG clusters.

1 - Configuration

Choose etcd cluster size based on requirements, provide reliable access.

Before deployment, define etcd cluster in config inventory. Typical choices:

  • One Node: No HA, suitable for dev, test, demo, or standalone deployments using external S3 backup for PITR
  • Three Nodes: Basic HA, tolerates 1 node failure, suitable for small-medium prod
  • Five Nodes: Better HA, tolerates 2 node failures, suitable for large prod

Even-numbered clusters don’t make sense; 5+ node clusters uncommon. Typical configs: single, 3-node, 5-node.

Cluster SizeQuorumFault ToleranceUse Case
1 node10Dev, test, demo
3 nodes21Small-medium prod
5 nodes32Large prod
7 nodes43Special HA requirements

One Node

Define singleton etcd instance in Pigsty—single line of config:

etcd: { hosts: { 10.10.10.10: { etcd_seq: 1 } }, vars: { etcd_cluster: etcd } }

All single-node config templates include this line. Placeholder IP 10.10.10.10 replaced with current admin node’s IP.

Only required params: etcd_seq and etcd_cluster—uniquely identify each etcd instance.


Three Nodes

Most common config: 3-node etcd cluster tolerates 1 node failure, suitable for small-medium prod.

Example: Pigsty’s 3-node templates trio and safe use 3-node etcd:

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }  # etcd_seq (instance number) required
    10.10.10.11: { etcd_seq: 2 }  # positive integers, sequential from 0 or 1
    10.10.10.12: { etcd_seq: 3 }  # immutable for life, never recycled
  vars: # cluster-level params
    etcd_cluster: etcd    # default cluster name: 'etcd', don't change unless deploying multiple etcd clusters
    etcd_safeguard: false # enable safeguard? Enable after prod init to prevent accidental deletion
    etcd_clean: true      # force remove existing during init? Enable for testing for true idempotency

Five Nodes

5-node cluster tolerates 2 node failures, suitable for large prod.

Example: Pigsty’s prod sim template prod uses 5-node etcd:

etcd:
  hosts:
    10.10.10.21 : { etcd_seq: 1 }
    10.10.10.22 : { etcd_seq: 2 }
    10.10.10.23 : { etcd_seq: 3 }
    10.10.10.24 : { etcd_seq: 4 }
    10.10.10.25 : { etcd_seq: 5 }
  vars: { etcd_cluster: etcd    }

Services Using etcd

Services using etcd in Pigsty:

ServicePurposeConfig File
PatroniPG HA, stores cluster state and config/pg/bin/patroni.yml
VIP-ManagerBinds L2 VIP on PG clusters/etc/default/vip-manager

When etcd cluster membership changes permanently, reload related service configs to ensure correct access.

Update Patroni’s etcd endpoint ref:

./pgsql.yml -t pg_conf                            # regenerate patroni config
ansible all -f 1 -b -a 'systemctl reload patroni' # reload patroni config

Update VIP-Manager’s etcd endpoint ref (only for PGSQL L2 VIP):

./pgsql.yml -t pg_vip_config                           # regenerate vip-manager config
ansible all -f 1 -b -a 'systemctl restart vip-manager' # restart vip-manager

RBAC Authentication Config

v4.0 enables etcd RBAC auth by default. Related params:

ParameterDescriptionDefault
etcd_root_passwordetcd root passwordEtcd.Root
pg_etcd_passwordPatroni’s password for etcdEmpty (uses cluster name)

Prod recommendations:

all:
  vars:
    etcd_root_password: 'YourSecureEtcdPassword'  # change default

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }
    10.10.10.11: { etcd_seq: 2 }
    10.10.10.12: { etcd_seq: 3 }
  vars:
    etcd_cluster: etcd
    etcd_safeguard: true    # enable safeguard for production

Filesystem Layout

Module creates these directories/files on target hosts:

PathPurposePermissions
/etc/etcd/Config dir0750, etcd:etcd
/etc/etcd/etcd.confMain config file0644, etcd:etcd
/etc/etcd/etcd.passRoot password file0640, root:etcd
/etc/etcd/ca.crtCA cert0644, etcd:etcd
/etc/etcd/server.crtServer cert0644, etcd:etcd
/etc/etcd/server.keyServer private key0600, etcd:etcd
/var/lib/etcd/Backup data dir0770, etcd:etcd
/data/etcd/Main data dir (configurable)0700, etcd:etcd
/etc/profile.d/etcdctl.shClient env vars0755, root:root
/etc/systemd/system/etcd.serviceSystemd service0644, root:root



2 - Parameters

ETCD module provides 13 configuration parameters for fine-grained control over cluster behavior.

The ETCD module has 13 parameters, divided into two sections:

  • ETCD: 10 parameters for etcd cluster deployment and configuration
  • ETCD_REMOVE: 3 parameters for controlling etcd cluster removal

Parameter Overview

The ETCD parameter group is used for etcd cluster deployment and configuration, including instance identification, cluster name, data directory, ports, and authentication password.

ParameterTypeLevelDescription
etcd_seqintIetcd instance identifier, REQUIRED
etcd_clusterstringCetcd cluster name, fixed to etcd by default
etcd_learnerboolI/Ainitialize etcd instance as learner?
etcd_datapathCetcd data directory, /data/etcd by default
etcd_portportCetcd client port, 2379 by default
etcd_peer_portportCetcd peer port, 2380 by default
etcd_initenumCetcd initial cluster state, new or existing
etcd_election_timeoutintCetcd election timeout, 1000ms by default
etcd_heartbeat_intervalintCetcd heartbeat interval, 100ms by default
etcd_root_passwordpasswordGetcd root user password for RBAC authentication

The ETCD_REMOVE parameter group controls etcd cluster removal behavior, including safeguard protection, data cleanup, and package uninstallation.

ParameterTypeLevelDescription
etcd_safeguardboolG/C/Asafeguard to prevent purging running etcd instances?
etcd_rm_databoolG/C/Aremove etcd data during removal? default is true
etcd_rm_pkgboolG/C/Auninstall etcd packages during removal? default is false

ETCD

This section contains parameters for the etcd role, which are used by the etcd.yml playbook.

Parameters are defined in roles/etcd/defaults/main.yml

#etcd_seq: 1                      # etcd instance identifier, explicitly required
etcd_cluster: etcd                # etcd cluster & group name, etcd by default
etcd_learner: false               # run etcd instance as learner? default is false
etcd_data: /data/etcd             # etcd data directory, /data/etcd by default
etcd_port: 2379                   # etcd client port, 2379 by default
etcd_peer_port: 2380              # etcd peer port, 2380 by default
etcd_init: new                    # etcd initial cluster state, new or existing
etcd_election_timeout: 1000       # etcd election timeout, 1000ms by default
etcd_heartbeat_interval: 100      # etcd heartbeat interval, 100ms by default
etcd_root_password: Etcd.Root     # etcd root user password for RBAC authentication (please change!)

etcd_seq

Parameter: etcd_seq, Type: int, Level: I

etcd instance identifier. This is a required parameter—you must assign a unique identifier to each etcd instance.

Here is an example of a 3-node etcd cluster with identifiers 1 through 3:

etcd: # dcs service for postgres/patroni ha consensus
  hosts:  # 1 node for testing, 3 or 5 for production
    10.10.10.10: { etcd_seq: 1 }  # etcd_seq required
    10.10.10.11: { etcd_seq: 2 }  # assign from 1 ~ n
    10.10.10.12: { etcd_seq: 3 }  # use odd numbers
  vars: # cluster level parameter override roles/etcd
    etcd_cluster: etcd  # mark etcd cluster name etcd
    etcd_safeguard: false # safeguard against purging

etcd_cluster

Parameter: etcd_cluster, Type: string, Level: C

etcd cluster & group name, default value is the hard-coded etcd.

You can modify this parameter when you want to deploy an additional etcd cluster for backup purposes.

etcd_learner

Parameter: etcd_learner, Type: bool, Level: I/A

Initialize etcd instance as learner? Default value is false.

When set to true, the etcd instance will be initialized as a learner, meaning it cannot participate in voting elections within the etcd cluster.

Use Cases:

  • Cluster Expansion: When adding new members to an existing cluster, using learner mode prevents affecting cluster quorum before data synchronization completes
  • Safe Migration: In rolling upgrade or migration scenarios, join as a learner first, then promote after confirming data synchronization

Workflow:

  1. Set etcd_learner: true to initialize the new member as a learner
  2. Wait for data synchronization to complete (check with etcdctl endpoint status)
  3. Use etcdctl member promote <member_id> to promote it to a full member

etcd_data

Parameter: etcd_data, Type: path, Level: C

etcd data directory, default is /data/etcd.

etcd_port

Parameter: etcd_port, Type: port, Level: C

etcd client port, default is 2379.

etcd_peer_port

Parameter: etcd_peer_port, Type: port, Level: C

etcd peer port, default is 2380.

etcd_init

Parameter: etcd_init, Type: enum, Level: C

etcd initial cluster state, can be new or existing, default value: new.

Option Values:

ValueDescriptionUse Case
newCreate a new etcd clusterInitial deployment, cluster rebuild
existingJoin an existing etcd clusterCluster expansion, adding new members

Important Notes:

Usage Examples:

# Create new cluster (default behavior)
./etcd.yml

# Add new member to existing cluster
./etcd.yml -l <new_ip> -e etcd_init=existing

# Or use the convenience script (automatically sets etcd_init=existing)
bin/etcd-add <new_ip>

etcd_election_timeout

Parameter: etcd_election_timeout, Type: int, Level: C

etcd election timeout, default is 1000 (milliseconds), i.e., 1 second.

etcd_heartbeat_interval

Parameter: etcd_heartbeat_interval, Type: int, Level: C

etcd heartbeat interval, default is 100 (milliseconds).

etcd_root_password

Parameter: etcd_root_password, Type: password, Level: G

etcd root user password for RBAC authentication, default value is Etcd.Root.

Pigsty v4.0 enables etcd RBAC (Role-Based Access Control) authentication by default. During cluster initialization, the etcd_auth task automatically creates the root user and enables authentication.

Password Storage Location:

  • Password is stored in /etc/etcd/etcd.pass file
  • File permissions are 0640 (owned by root, readable by etcd group)
  • The etcdctl environment script /etc/profile.d/etcdctl.sh automatically reads this file

Integration with Other Components:

  • Patroni uses the pg_etcd_password parameter to configure the password for connecting to etcd
  • If pg_etcd_password is empty, Patroni will use the cluster name as password (not recommended)
  • VIP-Manager also requires the same authentication credentials to connect to etcd

Security Recommendations:


ETCD_REMOVE

This section contains parameters for the etcd_remove role, which are action flags used by the etcd-rm.yml playbook.

Parameters are defined in roles/etcd_remove/defaults/main.yml

etcd_safeguard: false             # prevent purging running etcd instances?
etcd_rm_data: true                # remove etcd data and config files during removal?
etcd_rm_pkg: false                # uninstall etcd packages during removal?

etcd_safeguard

Parameter: etcd_safeguard, Type: bool, Level: G/C/A

Safeguard to prevent purging running etcd instances? Default value is false.

When enabled, the etcd-rm.yml playbook will abort when detecting running etcd instances, preventing accidental deletion of active etcd clusters.

Recommended Settings:

EnvironmentRecommendedDescription
Dev/TestfalseConvenient for rapid rebuilding and testing
ProductiontruePrevents service interruption from accidental operations

In emergencies, you can override the configuration with command-line parameters:

./etcd-rm.yml -e etcd_safeguard=false

etcd_rm_data

Parameter: etcd_rm_data, Type: bool, Level: G/C/A

Remove etcd data and configuration files during removal? Default value is true.

When enabled, the etcd-rm.yml playbook will delete the following contents when removing a cluster or member:

  • /etc/etcd/ - Configuration directory (including certificates and password files)
  • /var/lib/etcd/ - Alternate data directory
  • {{ etcd_data }} - Primary data directory (default /data/etcd)
  • {{ systemd_dir }}/etcd.service - Systemd service unit file
  • /etc/profile.d/etcdctl.sh - Client environment script
  • /etc/vector/etcd.yaml - Vector log collection config

Use Cases:

ScenarioRecommendedDescription
Complete removaltrue (default)Full cleanup, free disk space
Stop service onlyfalsePreserve data for troubleshooting or recovery
# Stop service only, preserve data
./etcd-rm.yml -e etcd_rm_data=false

etcd_rm_pkg

Parameter: etcd_rm_pkg, Type: bool, Level: G/C/A

Uninstall etcd packages during removal? Default value is false.

When enabled, the etcd-rm.yml playbook will uninstall etcd packages when removing a cluster or member.

Use Cases:

ScenarioRecommendedDescription
Normal removalfalse (default)Keep packages for quick redeployment
Complete cleanuptrueFull uninstall, save disk space
# Uninstall packages during removal
./etcd-rm.yml -e etcd_rm_pkg=true

3 - Administration

etcd cluster management SOP: create, destroy, scale, config, and RBAC.

Common etcd admin SOPs:

For more, refer to FAQ: ETCD.


Create Cluster

Define etcd cluster in config inventory:

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }
    10.10.10.11: { etcd_seq: 2 }
    10.10.10.12: { etcd_seq: 3 }
  vars: { etcd_cluster: etcd }

Run etcd.yml playbook:

./etcd.yml  # initialize etcd cluster

For prod etcd clusters, enable safeguard etcd_safeguard to prevent accidental deletion.


Destroy Cluster

Use dedicated etcd-rm.yml playbook to destroy etcd cluster. Use caution!

./etcd-rm.yml                         # remove entire etcd cluster
./etcd-rm.yml -e etcd_safeguard=false # override safeguard

Or use utility script:

bin/etcd-rm                           # remove entire etcd cluster

Removal playbook respects etcd_safeguard. If true, playbook aborts to prevent accidental deletion.


CLI Environment

Uses etcd v3 API by default (v2 removed in v3.6+). Pigsty auto-configures env script /etc/profile.d/etcdctl.sh on etcd nodes, loaded on login.

Example client env config:

alias e="etcdctl"
alias em="etcdctl member"
export ETCDCTL_ENDPOINTS=https://10.10.10.10:2379
export ETCDCTL_CACERT=/etc/etcd/ca.crt
export ETCDCTL_CERT=/etc/etcd/server.crt
export ETCDCTL_KEY=/etc/etcd/server.key

v4.0 enables RBAC auth by default—user auth required:

export ETCDCTL_USER="root:$(cat /etc/etcd/etcd.pass)"

After configuring client env, run etcd CRUD ops:

e put a 10 ; e get a; e del a   # basic KV ops
e member list                    # list cluster members
e endpoint health                # check endpoint health
e endpoint status                # view endpoint status

RBAC Authentication

v4.0 enables etcd RBAC auth by default. During cluster init, etcd_auth task auto-creates root user and enables auth.

Root user password set by etcd_root_password, default: Etcd.Root. Stored in /etc/etcd/etcd.pass with 0640 perms (root-owned, etcd-group readable).

Strongly recommended to change default password in prod:

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }
    10.10.10.11: { etcd_seq: 2 }
    10.10.10.12: { etcd_seq: 3 }
  vars:
    etcd_cluster: etcd
    etcd_root_password: 'YourSecurePassword'  # change default

Client auth methods:

# Method 1: env vars (recommended, auto-configured in /etc/profile.d/etcdctl.sh)
export ETCDCTL_USER="root:$(cat /etc/etcd/etcd.pass)"

# Method 2: command line
etcdctl --user root:YourSecurePassword member list

Patroni and etcd auth:

Patroni uses pg_etcd_password to configure etcd connection password. If empty, Patroni uses cluster name as password (not recommended). Configure separate etcd password per PG cluster in prod.


Reload Config

If etcd cluster membership changes (add/remove members), refresh etcd service endpoint references. These etcd refs in Pigsty need updates:

Config LocationConfig FileUpdate Method
etcd member config/etc/etcd/etcd.conf./etcd.yml -t etcd_conf
etcdctl env vars/etc/profile.d/etcdctl.sh./etcd.yml -t etcd_config
Patroni DCS config/pg/bin/patroni.yml./pgsql.yml -t pg_conf
VIP-Manager config/etc/default/vip-manager./pgsql.yml -t pg_vip_config

Refresh etcd member config:

./etcd.yml -t etcd_conf                           # refresh /etc/etcd/etcd.conf
ansible etcd -f 1 -b -a 'systemctl restart etcd'  # optional: restart etcd instances

Refresh etcdctl client env:

./etcd.yml -t etcd_config                         # refresh /etc/profile.d/etcdctl.sh

Update Patroni DCS endpoint config:

./pgsql.yml -t pg_conf                            # regenerate patroni config
ansible all -f 1 -b -a 'systemctl reload patroni' # reload patroni config

Update VIP-Manager endpoint config (only for PGSQL L2 VIP):

./pgsql.yml -t pg_vip_config                           # regenerate vip-manager config
ansible all -f 1 -b -a 'systemctl restart vip-manager' # restart vip-manager

Add Member

ETCD Reference: Add a member

Use bin/etcd-add script to add new members to existing etcd cluster:

# First add new member definition to config inventory, then:
bin/etcd-add <ip>              # add single new member
bin/etcd-add <ip1> <ip2> ...   # add multiple new members

Script auto-performs:

  • Validates IP address validity
  • Executes etcd.yml playbook (auto-sets etcd_init=existing)
  • Provides safety warnings and countdown
  • Prompts config refresh commands after completion

Manual: Step-by-Step

Add new member to existing etcd cluster:

  1. Update config inventory: Add new instance to etcd group
  2. Notify cluster: Run etcdctl member add (optional, playbook auto-does this)
  3. Initialize new member: Run playbook with etcd_init=existing parameter
  4. Promote member: Promote learner to full member (optional, required when using etcd_learner=true)
  5. Reload config: Update etcd endpoint references for all clients
# After config inventory update, initialize new member
./etcd.yml -l <new_ins_ip> -e etcd_init=existing

# If using learner mode, manually promote
etcdctl member promote <new_ins_server_id>
Detailed: Add member to etcd cluster

Detailed steps. Start from single-instance etcd cluster:

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 } # <--- only existing instance in cluster
    10.10.10.11: { etcd_seq: 2 } # <--- add this new member to inventory
  vars: { etcd_cluster: etcd }

Add new member using utility script (recommended):

$ bin/etcd-add 10.10.10.11

Or manual. First use etcdctl member add to announce new learner instance etcd-2 to existing etcd cluster:

$ etcdctl member add etcd-2 --learner=true --peer-urls=https://10.10.10.11:2380
Member 33631ba6ced84cf8 added to cluster 6646fbcf5debc68f

ETCD_NAME="etcd-2"
ETCD_INITIAL_CLUSTER="etcd-2=https://10.10.10.11:2380,etcd-1=https://10.10.10.10:2380"
ETCD_INITIAL_ADVERTISE_PEER_URLS="https://10.10.10.11:2380"
ETCD_INITIAL_CLUSTER_STATE="existing"

Check member list with etcdctl member list (or em list), see unstarted new member:

33631ba6ced84cf8, unstarted, , https://10.10.10.11:2380, , true       # unstarted new member here
429ee12c7fbab5c1, started, etcd-1, https://10.10.10.10:2380, https://10.10.10.10:2379, false

Next, use etcd.yml playbook to initialize new etcd instance etcd-2. After completion, new member has started:

$ ./etcd.yml -l 10.10.10.11 -e etcd_init=existing    # must add existing parameter
...
33631ba6ced84cf8, started, etcd-2, https://10.10.10.11:2380, https://10.10.10.11:2379, true
429ee12c7fbab5c1, started, etcd-1, https://10.10.10.10:2380, https://10.10.10.10:2379, false

After new member initialized and running stably, promote from learner to follower:

$ etcdctl member promote 33631ba6ced84cf8   # promote learner to follower
Member 33631ba6ced84cf8 promoted in cluster 6646fbcf5debc68f

$ em list                # check again, new member promoted to full member
33631ba6ced84cf8, started, etcd-2, https://10.10.10.11:2380, https://10.10.10.11:2379, false
429ee12c7fbab5c1, started, etcd-1, https://10.10.10.10:2380, https://10.10.10.10:2379, false

New member added. Don’t forget to reload config so all clients know new member.

Repeat steps to add more members. Prod environments need at least 3 members.


Remove Member

Use bin/etcd-rm script to remove members from etcd cluster:

bin/etcd-rm <ip>              # remove specified member
bin/etcd-rm <ip1> <ip2> ...   # remove multiple members
bin/etcd-rm                   # remove entire etcd cluster

Script auto-performs:

  • Gracefully removes members from cluster
  • Stops and disables etcd service
  • Cleans up data and config files
  • Deregisters from monitoring system

Manual: Step-by-Step

Remove member instance from etcd cluster:

  1. Remove from config inventory: Comment out or delete instance, and reload config
  2. Kick from cluster: Use etcdctl member remove command
  3. Clean up instance: Use etcd-rm.yml playbook to clean up
# Use dedicated removal playbook (recommended)
./etcd-rm.yml -l <ip>

# Or manual
etcdctl member remove <server_id>      # kick from cluster
./etcd-rm.yml -l <ip>                  # clean up instance
Detailed: Remove member from etcd cluster

Example: 3-node etcd cluster, remove instance 3.

Method 1: Utility script (recommended)

$ bin/etcd-rm 10.10.10.12

Script auto-completes all operations: remove from cluster, stop service, clean up data.

Method 2: Manual

First, refresh config by commenting out member to delete, then reload config so all clients stop using this instance.

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }
    10.10.10.11: { etcd_seq: 2 }
    # 10.10.10.12: { etcd_seq: 3 }   # <---- comment out this member
  vars: { etcd_cluster: etcd }

Then use removal playbook:

$ ./etcd-rm.yml -l 10.10.10.12

Playbook auto-executes:

  1. Get member list, find corresponding member ID
  2. Execute etcdctl member remove to kick from cluster
  3. Stop etcd service
  4. Clean up data and config files

If manual:

$ etcdctl member list
429ee12c7fbab5c1, started, etcd-1, https://10.10.10.10:2380, https://10.10.10.10:2379, false
33631ba6ced84cf8, started, etcd-2, https://10.10.10.11:2380, https://10.10.10.11:2379, false
93fcf23b220473fb, started, etcd-3, https://10.10.10.12:2380, https://10.10.10.12:2379, false  # <--- remove this

$ etcdctl member remove 93fcf23b220473fb # kick from cluster
Member 93fcf23b220473fb removed from cluster 6646fbcf5debc68f

After execution, permanently remove from config inventory. Member removal complete.

Repeat to remove more members. Combined with Add Member, perform rolling upgrades and migrations of etcd cluster.


Utility Scripts

v3.6+ provides utility scripts to simplify etcd cluster scaling:

bin/etcd-add

Add new members to existing etcd cluster:

bin/etcd-add <ip>              # add single new member
bin/etcd-add <ip1> <ip2> ...   # add multiple new members

Script features:

  • Validates IP addresses in config inventory
  • Auto-sets etcd_init=existing parameter
  • Executes etcd.yml playbook to complete member addition
  • Prompts config refresh commands after completion

bin/etcd-rm

Remove members or entire cluster from etcd:

bin/etcd-rm <ip>              # remove specified member
bin/etcd-rm <ip1> <ip2> ...   # remove multiple members
bin/etcd-rm                   # remove entire etcd cluster

Script features:

  • Provides safety warnings and confirmation countdown
  • Auto-executes etcd-rm.yml playbook
  • Gracefully removes members from cluster
  • Cleans up data and config files

4 - Playbook

Manage etcd clusters with Ansible playbooks and quick command reference.

The ETCD module provides two core playbooks: etcd.yml for installing and configuring etcd clusters, and etcd-rm.yml for removing etcd clusters or members.


etcd.yml

Playbook source: etcd.yml

This playbook installs and configures an etcd cluster on the hardcoded etcd group, then launches the etcd service.

The following subtasks are available in etcd.yml:

  • etcd_assert : Validate etcd identity parameters (etcd_seq must be defined as a non-negative integer)
  • etcd_install : Install etcd packages
  • etcd_dir : Create etcd data and configuration directories
  • etcd_config : Generate etcd configuration
    • etcd_conf : Generate etcd main config file /etc/etcd/etcd.conf
    • etcd_cert : Generate etcd TLS certificates (CA, server cert, private key)
  • etcd_member : Add new member to existing cluster (only runs when etcd_init=existing)
  • etcd_launch : Launch etcd service
  • etcd_auth : Enable RBAC authentication (create root user and enable auth)
  • etcd_register : Register etcd to VictoriaMetrics/Prometheus monitoring

etcd-rm.yml

Playbook source: etcd-rm.yml

A dedicated playbook for removing etcd clusters or individual members. The following subtasks are available in etcd-rm.yml:

  • etcd_safeguard : Check safeguard and abort if enabled
  • etcd_pause : Pause for 3 seconds, allowing user to abort with Ctrl-C
  • etcd_deregister : Remove etcd registration from VictoriaMetrics monitoring targets
  • etcd_leave : Try graceful leaving etcd cluster before purge
  • etcd_svc : Stop and disable etcd service with systemd
  • etcd_data : Remove etcd data (disable with etcd_rm_data=false)
  • etcd_pkg : Uninstall etcd packages (enable with etcd_rm_pkg=true)

The removal playbook uses the etcd_remove role with the following configurable parameters:

  • etcd_safeguard: Prevents accidental removal when set to true
  • etcd_rm_data: Controls whether ETCD data is deleted (default: true)
  • etcd_rm_pkg: Controls whether ETCD packages are uninstalled (default: false)

Demo

asciicast


Cheatsheet

Etcd Installation & Configuration:

./etcd.yml                                      # Initialize etcd cluster
./etcd.yml -t etcd_launch                       # Restart entire etcd cluster
./etcd.yml -t etcd_conf                         # Refresh /etc/etcd/etcd.conf with latest state
./etcd.yml -t etcd_cert                         # Regenerate etcd TLS certificates
./etcd.yml -l 10.10.10.12 -e etcd_init=existing # Scale out: add new member to existing cluster

Etcd Removal & Cleanup:

./etcd-rm.yml                                   # Remove entire etcd cluster
./etcd-rm.yml -l 10.10.10.12                    # Remove single etcd member
./etcd-rm.yml -e etcd_safeguard=false           # Override safeguard to force removal
./etcd-rm.yml -e etcd_rm_data=false             # Stop service only, preserve data
./etcd-rm.yml -e etcd_rm_pkg=true               # Also uninstall etcd packages

Convenience Scripts:

bin/etcd-add <ip>                               # Add new member to existing cluster (recommended)
bin/etcd-rm <ip>                                # Remove specific member from cluster (recommended)
bin/etcd-rm                                     # Remove entire etcd cluster

Safeguard

To prevent accidental deletion, Pigsty’s ETCD module provides a safeguard mechanism controlled by the etcd_safeguard parameter, which defaults to false (safeguard disabled).

For production etcd clusters that have been initialized, it’s recommended to enable the safeguard to prevent accidental deletion of existing etcd instances:

etcd:
  hosts:
    10.10.10.10: { etcd_seq: 1 }
    10.10.10.11: { etcd_seq: 2 }
    10.10.10.12: { etcd_seq: 3 }
  vars:
    etcd_cluster: etcd
    etcd_safeguard: true  # Enable safeguard protection

When etcd_safeguard is set to true, the etcd-rm.yml playbook will detect running etcd instances and abort to prevent accidental deletion. You can override this behavior using command-line parameters:

./etcd-rm.yml -e etcd_safeguard=false  # Force override safeguard

Unless you clearly understand what you’re doing, we do not recommend arbitrarily removing etcd clusters.

5 - Monitoring

etcd monitoring dashboards, metrics, and alert rules.

Dashboards

ETCD module provides one monitoring dashboard: Etcd Overview.

ETCD Overview Dashboard

ETCD Overview: Overview of ETCD cluster

Dashboard provides key ETCD status info. Notable: ETCD Aliveness—shows overall etcd cluster service status.

Red bands = instance downtime; blue-gray below = cluster unavailable.

etcd-overview.jpg


Alert Rules

Pigsty provides 5 preset alert rules for etcd, defined in files/prometheus/rules/etcd.yml:

  • EtcdServerDown: etcd node down, CRIT alert
  • EtcdNoLeader: etcd cluster no leader, CRIT alert
  • EtcdQuotaFull: etcd quota > 90%, WARN alert
  • EtcdNetworkPeerRTSlow: etcd network latency slow, INFO alert
  • EtcdWalFsyncSlow: etcd disk fsync slow, INFO alert
#==============================================================#
#                         Aliveness                            #
#==============================================================#
# etcd server instance down
- alert: EtcdServerDown
  expr: etcd_up < 1
  for: 1m
  labels: { level: 0, severity: CRIT, category: etcd }
  annotations:
    summary: "CRIT EtcdServerDown {{ $labels.ins }}@{{ $labels.instance }}"
    description: |
      etcd_up[ins={{ $labels.ins }}, instance={{ $labels.instance }}] = {{ $value }} < 1
      https://demo.pigsty.io/d/etcd-overview

#==============================================================#
#                         Error                                #
#==============================================================#
# Etcd no Leader triggers P0 alert immediately
# if dcs_failsafe mode not enabled, may cause global outage
- alert: EtcdNoLeader
  expr: min(etcd_server_has_leader) by (cls) < 1
  for: 15s
  labels: { level: 0, severity: CRIT, category: etcd }
  annotations:
    summary: "CRIT EtcdNoLeader: {{ $labels.cls }} {{ $value }}"
    description: |
      etcd_server_has_leader[cls={{ $labels.cls }}] = {{ $value }} < 1
      https://demo.pigsty.io/d/etcd-overview?from=now-5m&to=now&var-cls={{$labels.cls}}

#==============================================================#
#                        Saturation                            #
#==============================================================#
- alert: EtcdQuotaFull
  expr: etcd:cls:quota_usage > 0.90
  for: 1m
  labels: { level: 1, severity: WARN, category: etcd }
  annotations:
    summary: "WARN EtcdQuotaFull: {{ $labels.cls }}"
    description: |
      etcd:cls:quota_usage[cls={{ $labels.cls }}] = {{ $value | printf "%.3f" }} > 90%
      https://demo.pigsty.io/d/etcd-overview

#==============================================================#
#                         Latency                              #
#==============================================================#
# etcd network peer rt p95 > 200ms for 1m
- alert: EtcdNetworkPeerRTSlow
  expr: etcd:ins:network_peer_rt_p95_5m > 0.200
  for: 1m
  labels: { level: 2, severity: INFO, category: etcd }
  annotations:
    summary: "INFO EtcdNetworkPeerRTSlow: {{ $labels.cls }} {{ $labels.ins }}"
    description: |
      etcd:ins:network_peer_rt_p95_5m[cls={{ $labels.cls }}, ins={{ $labels.ins }}] = {{ $value }} > 200ms
      https://demo.pigsty.io/d/etcd-instance?from=now-10m&to=now&var-cls={{ $labels.cls }}
# Etcd wal fsync rt p95 > 50ms
- alert: EtcdWalFsyncSlow
  expr: etcd:ins:wal_fsync_rt_p95_5m > 0.050
  for: 1m
  labels: { level: 2, severity: INFO, category: etcd }
  annotations:
    summary: "INFO EtcdWalFsyncSlow: {{ $labels.cls }} {{ $labels.ins }}"
    description: |
      etcd:ins:wal_fsync_rt_p95_5m[cls={{ $labels.cls }}, ins={{ $labels.ins }}] = {{ $value }} > 50ms
      https://demo.pigsty.io/d/etcd-instance?from=now-10m&to=now&var-cls={{ $labels.cls }}

6 - Metrics

Complete monitoring metrics list provided by Pigsty ETCD module

The ETCD module has 177 available metrics.

Metric NameTypeLabelsDescription
etcd:ins:backend_commit_rt_p99_5mUnknowncls, ins, instance, job, ipN/A
etcd:ins:disk_fsync_rt_p99_5mUnknowncls, ins, instance, job, ipN/A
etcd:ins:network_peer_rt_p99_1mUnknowncls, To, ins, instance, job, ipN/A
etcd_cluster_versiongaugecls, cluster_version, ins, instance, job, ipRunning version. 1 = ‘cluster_version’ label with current version
etcd_debugging_auth_revisiongaugecls, ins, instance, job, ipCurrent auth store revision.
etcd_debugging_disk_backend_commit_rebalance_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_disk_backend_commit_rebalance_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_disk_backend_commit_rebalance_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_disk_backend_commit_spill_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_disk_backend_commit_spill_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_disk_backend_commit_spill_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_disk_backend_commit_write_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_disk_backend_commit_write_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_disk_backend_commit_write_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_lease_granted_totalcountercls, ins, instance, job, ipTotal granted leases.
etcd_debugging_lease_renewed_totalcountercls, ins, instance, job, ipRenewed leases seen by leader.
etcd_debugging_lease_revoked_totalcountercls, ins, instance, job, ipRevoked leases.
etcd_debugging_lease_ttl_total_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_lease_ttl_total_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_lease_ttl_total_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_compact_revisiongaugecls, ins, instance, job, ipLast compaction revision in store.
etcd_debugging_mvcc_current_revisiongaugecls, ins, instance, job, ipCurrent store revision.
etcd_debugging_mvcc_db_compaction_keys_totalcountercls, ins, instance, job, ipDB keys compacted.
etcd_debugging_mvcc_db_compaction_lastgaugecls, ins, instance, job, ipLast db compaction unix time. Resets to 0 on start.
etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_events_totalcountercls, ins, instance, job, ipEvents sent by this member.
etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_mvcc_keys_totalgaugecls, ins, instance, job, ipTotal keys.
etcd_debugging_mvcc_pending_events_totalgaugecls, ins, instance, job, ipPending events to send.
etcd_debugging_mvcc_range_totalcountercls, ins, instance, job, ipRanges seen by this member.
etcd_debugging_mvcc_slow_watcher_totalgaugecls, ins, instance, job, ipUnsynced slow watchers.
etcd_debugging_mvcc_total_put_size_in_bytesgaugecls, ins, instance, job, ipTotal put kv size seen by this member.
etcd_debugging_mvcc_watch_stream_totalgaugecls, ins, instance, job, ipWatch streams.
etcd_debugging_mvcc_watcher_totalgaugecls, ins, instance, job, ipWatchers.
etcd_debugging_server_lease_expired_totalcountercls, ins, instance, job, ipExpired leases.
etcd_debugging_snap_save_marshalling_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_snap_save_marshalling_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_snap_save_marshalling_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_snap_save_total_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_debugging_snap_save_total_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_debugging_snap_save_total_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_debugging_store_expires_totalcountercls, ins, instance, job, ipExpired keys.
etcd_debugging_store_reads_totalcountercls, action, ins, instance, job, ipReads (get/getRecursive) to this member.
etcd_debugging_store_watch_requests_totalcountercls, ins, instance, job, ipIncoming watch requests (new/reestablished).
etcd_debugging_store_watchersgaugecls, ins, instance, job, ipActive watchers.
etcd_debugging_store_writes_totalcountercls, action, ins, instance, job, ipWrites (set/compareAndDelete) to this member.
etcd_disk_backend_commit_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_disk_backend_commit_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_disk_backend_commit_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_disk_backend_defrag_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_disk_backend_defrag_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_disk_backend_defrag_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_disk_backend_snapshot_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_disk_backend_snapshot_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_disk_backend_snapshot_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_disk_defrag_inflightgaugecls, ins, instance, job, ipDefrag active. 1 = active, 0 = not.
etcd_disk_wal_fsync_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_disk_wal_fsync_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_disk_wal_fsync_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_disk_wal_write_bytes_totalgaugecls, ins, instance, job, ipWAL bytes written.
etcd_grpc_proxy_cache_hits_totalgaugecls, ins, instance, job, ipCache hits.
etcd_grpc_proxy_cache_keys_totalgaugecls, ins, instance, job, ipKeys/ranges cached.
etcd_grpc_proxy_cache_misses_totalgaugecls, ins, instance, job, ipCache misses.
etcd_grpc_proxy_events_coalescing_totalcountercls, ins, instance, job, ipEvents coalescing.
etcd_grpc_proxy_watchers_coalescing_totalgaugecls, ins, instance, job, ipCurrent watchers coalescing.
etcd_mvcc_db_open_read_transactionsgaugecls, ins, instance, job, ipOpen read transactions.
etcd_mvcc_db_total_size_in_bytesgaugecls, ins, instance, job, ipDB physical bytes allocated.
etcd_mvcc_db_total_size_in_use_in_bytesgaugecls, ins, instance, job, ipDB logical bytes in use.
etcd_mvcc_delete_totalcountercls, ins, instance, job, ipDeletes seen by this member.
etcd_mvcc_hash_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_mvcc_hash_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_mvcc_hash_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_mvcc_hash_rev_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_mvcc_hash_rev_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_mvcc_hash_rev_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_mvcc_put_totalcountercls, ins, instance, job, ipPuts seen by this member.
etcd_mvcc_range_totalcountercls, ins, instance, job, ipRanges seen by this member.
etcd_mvcc_txn_totalcountercls, ins, instance, job, ipTxns seen by this member.
etcd_network_active_peersgaugecls, ins, Local, instance, job, ip, RemoteActive peer connections.
etcd_network_client_grpc_received_bytes_totalcountercls, ins, instance, job, ipgRPC client bytes received.
etcd_network_client_grpc_sent_bytes_totalcountercls, ins, instance, job, ipgRPC client bytes sent.
etcd_network_peer_received_bytes_totalcountercls, ins, instance, job, ip, FromPeer bytes received.
etcd_network_peer_round_trip_time_seconds_bucketUnknowncls, To, ins, instance, job, le, ipN/A
etcd_network_peer_round_trip_time_seconds_countUnknowncls, To, ins, instance, job, ipN/A
etcd_network_peer_round_trip_time_seconds_sumUnknowncls, To, ins, instance, job, ipN/A
etcd_network_peer_sent_bytes_totalcountercls, To, ins, instance, job, ipPeer bytes sent.
etcd_server_apply_duration_seconds_bucketUnknowncls, version, ins, instance, job, le, success, ip, opN/A
etcd_server_apply_duration_seconds_countUnknowncls, version, ins, instance, job, success, ip, opN/A
etcd_server_apply_duration_seconds_sumUnknowncls, version, ins, instance, job, success, ip, opN/A
etcd_server_client_requests_totalcounterclient_api_version, cls, ins, instance, type, job, ipClient requests per version.
etcd_server_go_versiongaugecls, ins, instance, job, server_go_version, ipGo version running. 1 = ‘server_go_version’ label with current version.
etcd_server_has_leadergaugecls, ins, instance, job, ipLeader exists. 1 = exists, 0 = not.
etcd_server_health_failurescountercls, ins, instance, job, ipFailed health checks.
etcd_server_health_successcountercls, ins, instance, job, ipSuccessful health checks.
etcd_server_heartbeat_send_failures_totalcountercls, ins, instance, job, ipLeader heartbeat send failures (likely overloaded from slow disk).
etcd_server_idgaugecls, ins, instance, job, server_id, ipServer/member ID (hex). 1 = ‘server_id’ label with current ID.
etcd_server_is_leadergaugecls, ins, instance, job, ipMember is leader. 1 if is, 0 otherwise.
etcd_server_is_learnergaugecls, ins, instance, job, ipMember is learner. 1 if is, 0 otherwise.
etcd_server_leader_changes_seen_totalcountercls, ins, instance, job, ipLeader changes seen.
etcd_server_learner_promote_successescountercls, ins, instance, job, ipSuccessful learner promotions while this member is leader.
etcd_server_proposals_applied_totalgaugecls, ins, instance, job, ipConsensus proposals applied.
etcd_server_proposals_committed_totalgaugecls, ins, instance, job, ipConsensus proposals committed.
etcd_server_proposals_failed_totalcountercls, ins, instance, job, ipFailed proposals seen.
etcd_server_proposals_pendinggaugecls, ins, instance, job, ipPending proposals to commit.
etcd_server_quota_backend_bytesgaugecls, ins, instance, job, ipBackend storage quota bytes.
etcd_server_read_indexes_failed_totalcountercls, ins, instance, job, ipFailed read indexes seen.
etcd_server_slow_apply_totalcountercls, ins, instance, job, ipSlow apply requests (likely overloaded from slow disk).
etcd_server_slow_read_indexes_totalcountercls, ins, instance, job, ipPending read indexes not in sync with leader or timed out read index requests.
etcd_server_snapshot_apply_in_progress_totalgaugecls, ins, instance, job, ip1 if server applying incoming snapshot. 0 if none.
etcd_server_versiongaugecls, server_version, ins, instance, job, ipVersion running. 1 = ‘server_version’ label with current version.
etcd_snap_db_fsync_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_snap_db_fsync_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_snap_db_fsync_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_snap_db_save_total_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_snap_db_save_total_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_snap_db_save_total_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_snap_fsync_duration_seconds_bucketUnknowncls, ins, instance, job, le, ipN/A
etcd_snap_fsync_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
etcd_snap_fsync_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
etcd_upUnknowncls, ins, instance, job, ipN/A
go_gc_duration_secondssummarycls, ins, instance, job, quantile, ipGC pause duration summary.
go_gc_duration_seconds_countUnknowncls, ins, instance, job, ipN/A
go_gc_duration_seconds_sumUnknowncls, ins, instance, job, ipN/A
go_goroutinesgaugecls, ins, instance, job, ipGoroutines.
go_infogaugecls, version, ins, instance, job, ipGo environment info.
go_memstats_alloc_bytesgaugecls, ins, instance, job, ipBytes allocated and in use.
go_memstats_alloc_bytes_totalcountercls, ins, instance, job, ipBytes allocated, even if freed.
go_memstats_buck_hash_sys_bytesgaugecls, ins, instance, job, ipBytes used by profiling bucket hash table.
go_memstats_frees_totalcountercls, ins, instance, job, ipFrees.
go_memstats_gc_cpu_fractiongaugecls, ins, instance, job, ipGC CPU fraction since program started.
go_memstats_gc_sys_bytesgaugecls, ins, instance, job, ipBytes used for GC system metadata.
go_memstats_heap_alloc_bytesgaugecls, ins, instance, job, ipHeap bytes allocated and in use.
go_memstats_heap_idle_bytesgaugecls, ins, instance, job, ipHeap bytes waiting to be used.
go_memstats_heap_inuse_bytesgaugecls, ins, instance, job, ipHeap bytes in use.
go_memstats_heap_objectsgaugecls, ins, instance, job, ipAllocated objects.
go_memstats_heap_released_bytesgaugecls, ins, instance, job, ipHeap bytes released to OS.
go_memstats_heap_sys_bytesgaugecls, ins, instance, job, ipHeap bytes obtained from system.
go_memstats_last_gc_time_secondsgaugecls, ins, instance, job, ipSeconds since 1970 of last GC.
go_memstats_lookups_totalcountercls, ins, instance, job, ipPointer lookups.
go_memstats_mallocs_totalcountercls, ins, instance, job, ipMallocs.
go_memstats_mcache_inuse_bytesgaugecls, ins, instance, job, ipBytes in use by mcache structures.
go_memstats_mcache_sys_bytesgaugecls, ins, instance, job, ipBytes used for mcache structures from system.
go_memstats_mspan_inuse_bytesgaugecls, ins, instance, job, ipBytes in use by mspan structures.
go_memstats_mspan_sys_bytesgaugecls, ins, instance, job, ipBytes used for mspan structures from system.
go_memstats_next_gc_bytesgaugecls, ins, instance, job, ipHeap bytes when next GC will take place.
go_memstats_other_sys_bytesgaugecls, ins, instance, job, ipBytes used for other system allocations.
go_memstats_stack_inuse_bytesgaugecls, ins, instance, job, ipBytes in use by stack allocator.
go_memstats_stack_sys_bytesgaugecls, ins, instance, job, ipBytes obtained from system for stack allocator.
go_memstats_sys_bytesgaugecls, ins, instance, job, ipBytes obtained from system.
go_threadsgaugecls, ins, instance, job, ipOS threads created.
grpc_server_handled_totalcountercls, ins, instance, job, grpc_code, grpc_method, grpc_type, ip, grpc_serviceRPCs completed on server.
grpc_server_msg_received_totalcountercls, ins, instance, job, grpc_type, grpc_method, ip, grpc_serviceRPC stream messages received on server.
grpc_server_msg_sent_totalcountercls, ins, instance, job, grpc_type, grpc_method, ip, grpc_servicegRPC stream messages sent on server.
grpc_server_started_totalcountercls, ins, instance, job, grpc_type, grpc_method, ip, grpc_serviceRPCs started on server.
os_fd_limitgaugecls, ins, instance, job, ipFD limit.
os_fd_usedgaugecls, ins, instance, job, ipUsed FDs.
process_cpu_seconds_totalcountercls, ins, instance, job, ipUser + system CPU seconds.
process_max_fdsgaugecls, ins, instance, job, ipMax FDs.
process_open_fdsgaugecls, ins, instance, job, ipOpen FDs.
process_resident_memory_bytesgaugecls, ins, instance, job, ipResident memory bytes.
process_start_time_secondsgaugecls, ins, instance, job, ipStart time (unix epoch seconds).
process_virtual_memory_bytesgaugecls, ins, instance, job, ipVirtual memory bytes.
process_virtual_memory_max_bytesgaugecls, ins, instance, job, ipMax virtual memory bytes.
promhttp_metric_handler_requests_in_flightgaugecls, ins, instance, job, ipCurrent scrapes.
promhttp_metric_handler_requests_totalcountercls, ins, instance, job, ip, codeScrapes by HTTP status code.
scrape_duration_secondsUnknowncls, ins, instance, job, ipN/A
scrape_samples_post_metric_relabelingUnknowncls, ins, instance, job, ipN/A
scrape_samples_scrapedUnknowncls, ins, instance, job, ipN/A
scrape_series_addedUnknowncls, ins, instance, job, ipN/A
upUnknowncls, ins, instance, job, ipN/A

7 - FAQ

Frequently asked questions about Pigsty etcd module

What is etcd’s role in Pigsty?

etcd is a distributed, reliable key-value store for critical system data. Pigsty uses etcd as DCS (Distributed Config Store) service for Patroni, storing PG HA status.

Patroni uses etcd for: cluster failure detection, auto failover, primary-replica switchover, and cluster config management.

etcd is critical for PG HA. etcd’s availability and DR ensured through multiple distributed nodes.


What’s the appropriate etcd cluster size?

If more than half (including exactly half) of etcd instances unavailable, etcd cluster enters unavailable state—refuses service.

Example: 3-node cluster allows max 1 node failure while 2 others continue; 5-node cluster tolerates 2 node failures.

Note: Learner instances don’t count toward members—3-node cluster with 1 learner = 2 actual members, zero fault tolerance.

In prod, use odd number of instances. For prod, recommend 3-node or 5-node for reliability.


Impact of etcd unavailability?

If etcd cluster unavailable, affects PG control plane but not data plane—existing PG clusters continue running, but Patroni management ops fail.

During etcd failure: PG HA can’t auto failover, can’t use patronictl for PG management (config changes, manual failover, etc.).

Ansible playbooks unaffected by etcd failure: create DB, create user, refresh HBA/Service config. During etcd failure, operate PG clusters directly.

Note: Behavior applies to Patroni >=3.0 (Pigsty >=2.0). With older Patroni (<3.0, Pigsty 1.x), etcd/consul failure causes severe global impact:

All PG clusters demote: primaries → replicas, reject writes, etcd failure amplifies to global PG failure. Patroni 3.0 introduced DCS Failsafe—significantly improved.


What data does etcd store?

In Pigsty, etcd is PG HA only—no other config/state data.

PG HA component Patroni auto-generates and manages etcd data. If lost in etcd, Patroni auto-rebuilds.

Thus, by default, etcd in Pigsty = “stateless service”—destroyable and rebuildable, simplifies maintenance.

If using etcd for other purposes (K8s metadata, custom storage), backup etcd data yourself and restore after cluster recovery.


Recover from etcd failure?

Since etcd in Pigsty = PG HA only = “stateless service”—disposable, rebuildable. Failures? “restart” or “reset” to stop bleeding.

Restart etcd cluster:

./etcd.yml -t etcd_launch

Reset etcd cluster:

./etcd.yml

For custom etcd data: backup and restore after recovery.


Etcd maintenance considerations?

Simple answer: don’t fill up etcd.

Pigsty v2.6+ enables etcd auto-compaction and 16GB backend quota—usually fine.

etcd’s data model = each write generates new version.

Frequent writes (even few keys) = growing etcd DB size. At capacity limit, etcd rejects writes → PG HA breaks.

Pigsty’s default etcd config includes optimizations:

auto-compaction-mode: periodic      # periodic auto compaction
auto-compaction-retention: "24h"    # retain 24 hours history
quota-backend-bytes: 17179869184    # 16 GiB quota

More details: etcd official maintenance guide.


Enable etcd auto garbage collection?

Earlier Pigsty (v2.0 - v2.5)? Enable etcd auto-compaction in prod to avoid quota-based unavailability.

Edit etcd config template: roles/etcd/templates/etcd.conf.j2:

auto-compaction-mode: periodic
auto-compaction-retention: "24h"
quota-backend-bytes: 17179869184

Then set related PG clusters to maintenance mode and redeploy etcd with ./etcd.yml.

This increases default quota from 2 GiB → 16 GiB, retains last 24h writes—avoids infinite growth.


Where is PG HA data stored in etcd?

By default, Patroni uses pg_namespace prefix (default: /pg) for all metadata keys, followed by PG cluster name.

Example: PG cluster pg-meta stores metadata under /pg/pg-meta.

etcdctl get /pg/pg-meta --prefix

Sample data:

/pg/pg-meta/config
{"ttl":30,"loop_wait":10,"retry_timeout":10,"primary_start_timeout":10,"maximum_lag_on_failover":1048576,"maximum_lag_on_syncnode":-1,"primary_stop_timeout":30,"synchronous_mode":false,"synchronous_mode_strict":false,"failsafe_mode":true,"pg_version":16,"pg_cluster":"pg-meta","pg_shard":"pg-meta","pg_group":0,"postgresql":{"use_slots":true,"use_pg_rewind":true,"remove_data_directory_on_rewind_failure":true,"parameters":{"max_connections":100,"superuser_reserved_connections":10,"max_locks_per_transaction":200,"max_prepared_transactions":0,"track_commit_timestamp":"on","wal_level":"logical","wal_log_hints":"on","max_worker_processes":16,"max_wal_senders":50,"max_replication_slots":50,"password_encryption":"scram-sha-256","ssl":"on","ssl_cert_file":"/pg/cert/server.crt","ssl_key_file":"/pg/cert/server.key","ssl_ca_file":"/pg/cert/ca.crt","shared_buffers":"7969MB","maintenance_work_mem":"1993MB","work_mem":"79MB","max_parallel_workers":8,"max_parallel_maintenance_workers":2,"max_parallel_workers_per_gather":0,"hash_mem_multiplier":8.0,"huge_pages":"try","temp_file_limit":"7GB","vacuum_cost_delay":"20ms","vacuum_cost_limit":2000,"bgwriter_delay":"10ms","bgwriter_lru_maxpages":800,"bgwriter_lru_multiplier":5.0,"min_wal_size":"7GB","max_wal_size":"28GB","max_slot_wal_keep_size":"42GB","wal_buffers":"16MB","wal_writer_delay":"20ms","wal_writer_flush_after":"1MB","commit_delay":20,"commit_siblings":10,"checkpoint_timeout":"15min","checkpoint_completion_target":0.8,"archive_mode":"on","archive_timeout":300,"archive_command":"pgbackrest --stanza=pg-meta archive-push %p","max_standby_archive_delay":"10min","max_standby_streaming_delay":"3min","wal_receiver_status_interval":"1s","hot_standby_feedback":"on","wal_receiver_timeout":"60s","max_logical_replication_workers":8,"max_sync_workers_per_subscription":6,"random_page_cost":1.1,"effective_io_concurrency":1000,"effective_cache_size":"23907MB","default_statistics_target":200,"log_destination":"csvlog","logging_collector":"on","l...
ode=prefer"}}
/pg/pg-meta/failsafe
{"pg-meta-2":"http://10.10.10.11:8008/patroni","pg-meta-1":"http://10.10.10.10:8008/patroni"}
/pg/pg-meta/initialize
7418384210787662172
/pg/pg-meta/leader
pg-meta-1
/pg/pg-meta/members/pg-meta-1
{"conn_url":"postgres://10.10.10.10:5432/postgres","api_url":"http://10.10.10.10:8008/patroni","state":"running","role":"primary","version":"4.0.1","tags":{"clonefrom":true,"version":"16","spec":"8C.32G.125G","conf":"tiny.yml"},"xlog_location":184549376,"timeline":1}
/pg/pg-meta/members/pg-meta-2
{"conn_url":"postgres://10.10.10.11:5432/postgres","api_url":"http://10.10.10.11:8008/patroni","state":"running","role":"replica","version":"4.0.1","tags":{"clonefrom":true,"version":"16","spec":"8C.32G.125G","conf":"tiny.yml"},"xlog_location":184549376,"replication_state":"streaming","timeline":1}
/pg/pg-meta/status
{"optime":184549376,"slots":{"pg_meta_2":184549376,"pg_meta_1":184549376},"retain_slots":["pg_meta_1","pg_meta_2"]}

Use external existing etcd cluster?

Config inventory hardcodes etcd group—members used as DCS servers for PGSQL. Initialize with etcd.yml or assume external cluster exists.

To use external etcd: define as usual. Skip etcd.yml execution since cluster exists—no deployment needed.

Requirement: external etcd cluster certificate must use same CA as Pigsty—otherwise clients can’t use Pigsty’s self-signed certs.


Add new member to existing etcd cluster?

For detailed process, refer to Add member to etcd cluster

Recommended: Utility script

# First add new member to config inventory, then:
bin/etcd-add <ip>      # add single new member
bin/etcd-add <ip1>     # add multiple new members

Manual method:

etcdctl member add <etcd-?> --learner=true --peer-urls=https://<new_ins_ip>:2380 # announce new member
./etcd.yml -l <new_ins_ip> -e etcd_init=existing                                 # initialize new member
etcdctl member promote <new_ins_server_id>                                       # promote to full member

Recommend: add one new member at a time.


Remove member from existing etcd cluster?

For detailed process, refer to Remove member from etcd cluster

Recommended: Utility script

bin/etcd-rm <ip>              # remove specified member
bin/etcd-rm                   # remove entire etcd cluster

Manual method:

./etcd-rm.yml -l <ins_ip>                    # use dedicated removal playbook
etcdctl member remove <etcd_server_id>       # kick from cluster
./etcd-rm.yml -l <ins_ip>                    # clean up instance

Configure etcd RBAC authentication?

Pigsty v4.0 enables etcd RBAC auth by default. Root password set by etcd_root_password, default: Etcd.Root.

Prod recommendation: change default password

all:
  vars:
    etcd_root_password: 'YourSecurePassword'

Client auth:

# On etcd nodes, env vars auto-configured
source /etc/profile.d/etcdctl.sh
etcdctl member list

# Manual auth config
export ETCDCTL_USER="root:YourSecurePassword"
export ETCDCTL_CACERT=/etc/etcd/ca.crt
export ETCDCTL_CERT=/etc/etcd/server.crt
export ETCDCTL_KEY=/etc/etcd/server.key

More: RBAC Authentication.