with one click
local-cluster-manager
// Manage local multigres cluster components (multipooler, pgctld, multiorch, multigateway) - start/stop services, view logs, connect with psql, test S3 backups locally
// Manage local multigres cluster components (multipooler, pgctld, multiorch, multigateway) - start/stop services, view logs, connect with psql, test S3 backups locally
[HINT] Download the complete skill directory including SKILL.md and all related files
| name | Local Cluster Manager |
| description | Manage local multigres cluster components (multipooler, pgctld, multiorch, multigateway) - start/stop services, view logs, connect with psql, test S3 backups locally |
Manage local multigres cluster - both cluster-wide operations and individual components.
Invoke this skill when the user asks to:
Parse ./multigres_local/multigres.yaml once when this skill is first invoked and cache the cluster configuration in memory for the duration of the conversation. Use the cached data for all subsequent commands. Only re-parse if the user explicitly asks to "reload config" or if a command fails due to stale config.
Start entire cluster:
./bin/multigres cluster start
Stop entire cluster:
./bin/multigres cluster stop
Stop entire cluster and delete all cluster data:
./bin/multigres cluster stop --clean
Check cluster status:
./bin/multigres cluster status
Initialize new cluster:
./bin/multigres cluster init
Get all multipoolers from topology:
./bin/multigres getpoolers
Returns JSON with all multipoolers, their cells, service IDs, ports, and pooler directories.
Get detailed status for a specific multipooler:
./bin/multigres getpoolerstatus --cell <cell-name> --service-id <service-id>
Returns detailed status including:
pooler_type: 1 = PRIMARY, 2 = REPLICApostgres_role: "primary" or "standby"postgres_running: Whether PostgreSQL is runningwal_position: Current WAL positionconsensus_term: Current consensus termprimary_status: (for PRIMARY) connected followers and sync replication configreplication_status: (for REPLICA) replication lag and primary connection infoExample:
./bin/multigres getpoolerstatus --cell zone1 --service-id thhcdhbp
Check PostgreSQL recovery mode directly:
psql -h <pooler-dir>/pg_sockets -p <pg-port> -U postgres -d postgres -c "SELECT pg_is_in_recovery();"
Returns t (true) if in recovery/standby mode, f (false) if primary.
Test S3 backups using AWS S3. When the user wants to test S3 backups:
Configuration Caching: When S3 configuration values are first provided, cache them in memory for the duration of the conversation. Reuse these cached values for all subsequent S3 operations. Only re-prompt if:
Prompt for S3 configuration using AskUserQuestion (only if not already cached):
./.staging-aws or ~/.aws/credentials)s3://bucket-name/backups/)us-east-1)Check/source credentials:
# Check if AWS credentials are already set
env | grep AWS_
# If not, source the credentials file (path from user)
source <credentials-file-path>
# Verify credentials are now set
env | grep AWS_
IMPORTANT:
./bin/multigres cluster stop --clean
rm -rf multigres_local
./bin/multigres cluster init \
--backup-url=<s3-url-from-user> \
--region=<region-from-user>
Start cluster (use standard cluster start command)
Verify S3 configuration:
grep -r "aws_access_key_id\|aws_secret_access_key\|region\|repo1-s3" ./multigres_local/data/pooler_*/pgbackrest.conf
Should see AWS credentials and S3 configuration in all pgbackrest.conf files.
Create backup:
./bin/multigres cluster backup
List all backups:
./bin/multigres cluster list-backups
Restore from backup:
./bin/multigres cluster restore --backup-label <label>
Missing/expired credentials:
# Re-source credentials file
source <credentials-file-path>
# Verify they're set
env | grep AWS_ | wc -l # Should show 3+ environment variables
# Reinitialize cluster to pick up new credentials
./bin/multigres cluster stop --clean
rm -rf multigres_local
./bin/multigres cluster init --backup-url=<s3-url> --region=<region>
Check pgbackrest logs for errors:
# View recent errors
tail -100 ./multigres_local/data/pooler_*/pg_data/log/pgbackrest-*.log
# Follow logs in real-time
tail -f ./multigres_local/data/pooler_*/pg_data/log/pgbackrest-*.log
Verify S3 bucket access:
# Use AWS CLI to test bucket access (if installed)
aws s3 ls <s3-bucket-path> --region <region>
Start the observability stack (Grafana + Prometheus + Loki + Tempo) for metrics, traces, and logs visualization.
Start cluster with observability:
# 1. Start observability stack (separate terminal, runs in foreground)
demo/local/run-observability.sh
# 2. Start cluster with OTel export (separate terminal)
demo/local/multigres-with-otel.sh cluster start --config-path <config-path>
Generate traffic with pgbench:
Run pgbench init synchronously first, then start the workload in a background Agent so the user sees the output when it completes (do NOT use run_in_background on Bash — that hides output).
The local cluster runs one multigateway per cell (zone1/zone2/zone3 by default), each on its own pg-port (15432/15433/15434). Each gateway maintains independent in-memory state — its own query registry, consolidator, and connection pool.
That means a single-port -h localhost -p 15432 workload only exercises one gateway and leaves the other two idle. That hides per-instance bugs and means the per-gateway diagnostic pages on the other gateways stay empty.
To exercise all gateways, pass a libpq multi-host conninfo string with load_balance_hosts=random (PostgreSQL 16+; the pgbench and psql shipped with PG 17 honor it). Each new connection picks a gateway at random; with -c 9 clients you typically land ~3 connections per gateway. Connections are sticky for their lifetime, so distribution evens out across connections, not within.
Discover the gateway pg-ports from the cluster's cached config (every multigateway entry has its own pg-port) and build the conninfo string from there.
# Conninfo with all multigateway pg-ports — substitute the actual ports
# from the cluster config (default local layout shown).
CONNSTR='host=localhost,localhost,localhost port=15432,15433,15434 dbname=postgres user=postgres password=postgres load_balance_hosts=random'
# Step 1: Init (synchronous; init's single connection picks one gateway
# at random — either route ends up at the same primary postgres).
PGPASSWORD=postgres pgbench -i "$CONNSTR"
# Step 2: Workload (run in a background Agent). -c is a multiple of the
# gateway count for even distribution.
pgbench -c 9 -j 3 -T 300 -P 5 "$CONNSTR"
If the user only wants to drive a single gateway on purpose (e.g. reproducing a specific instance's bug), fall back to -h localhost -p <port> and call out the choice — don't silently single-target.
View telemetry:
Teardown (stop in this order to avoid OTel export errors):
# 1. Stop the cluster first
./bin/multigres cluster stop --config-path <config-path>
# 2. Stop the observability stack
docker rm -f multigres-observability
Full restart:
# Teardown
./bin/multigres cluster stop --config-path <config-path>
docker rm -f multigres-observability
# Start
demo/local/run-observability.sh # terminal 1
demo/local/multigres-with-otel.sh cluster start --config-path <config-path> # terminal 2
Observability ports:
| Service | Port |
|---|---|
| Grafana | 3000 |
| OTLP (HTTP) | 4318 |
| Prometheus | 9090 |
| Loki | 3100 |
| Tempo | 3200 |
Parse the config: Read ./multigres_local/multigres.yaml to discover available components and their IDs
Component ID mapping:
.provisioner-config.cells.<zone>.multipooler.service-idIf no ID provided: Use AskUserQuestion to let the user select which instance to operate on
Stop pgctld:
./bin/pgctld stop --pooler-dir <pooler-dir-from-config>
Start pgctld:
./bin/pgctld start --pooler-dir <pooler-dir-from-config>
Restart pgctld (as standby):
./bin/pgctld restart --pooler-dir <pooler-dir-from-config> --as-standby
Check pgctld status:
./bin/pgctld status --pooler-dir <pooler-dir-from-config>
View logs:
./multigres_local/logs/dbs/postgres/multipooler/[id].log./multigres_local/logs/dbs/postgres/pgctld/[id].log./multigres_local/logs/dbs/postgres/multiorch/[id].log./multigres_local/logs/dbs/postgres/multigateway/[id].log./multigres_local/data/pooler_[id]/pg_data/postgresql.logTail logs:
tail -f <log-path>
Connect to multipooler (via Unix socket):
psql -h <pooler-dir>/pg_sockets -p <pg-port> -U postgres -d postgres
Where:
.provisioner-config.cells.<zone>.multipooler.pooler-dir.provisioner-config.cells.<zone>.pgctld.pg-port<pooler-dir>/pg_sockets/.s.PGSQL.<pg-port>Example:
psql -h ./multigres_local/data/pooler_xf42rpl6/pg_sockets -p 25432 -U postgres -d postgres
Connect to multigateway (via TCP):
psql -h localhost -p <pg-port> -U postgres -d postgres
Where:
.provisioner-config.cells.<zone>.multigateway.pg-portExample:
psql -h localhost -p 15432 -U postgres -d postgres
Extract from YAML config at .provisioner-config.cells.<zone>.pgctld.pooler-dir
Cluster-wide:
User: "start the cluster"
./bin/multigres cluster startUser: "stop cluster"
./bin/multigres cluster stopUser: "cluster status"
./bin/multigres cluster statusUser: "show me all multipoolers" or "get poolers"
./bin/multigres getpoolersUser: "check if multipoolers are in recovery" or "check multipooler status"
./bin/multigres getpoolerstatus --cell <zone> --service-id <id> for eachUser: "check zone1 multipooler status"
./bin/multigres getpoolerstatus --cell zone1 --service-id <id>Observability:
User: "start cluster with otel" or "start cluster with observability"
demo/local/run-observability.sh (if not running)demo/local/multigres-with-otel.sh cluster start --config-path <path>User: "teardown everything" or "stop everything"
./bin/multigres cluster stop --config-path <path>docker rm -f multigres-observabilityUser: "restart everything" or "full restart"
User: "push traffic" or "generate load"
load_balance_hosts=random so traffic distributes across every multigateway instance — never default to single-port -h … -p … (that hides per-instance bugs and leaves the other gateways' diagnostic pages empty)-P 5 for progress-c as a multiple of the gateway count (e.g. -c 9 for 3 gateways) so each gateway gets at least a few sticky connectionsIndividual components:
User: "stop pgctld"
User: "restart pgctld xf42rpl6 as standby"
./bin/pgctld restart --pooler-dir /path/to/pooler_xf42rpl6 --as-standbyUser: "logs multipooler hm9hmxzm"
./multigres_local/logs/dbs/postgres/multipooler/hm9hmxzm.logUser: "tail pgctld"
User: "connect to multipooler zone1" or "psql multipooler xf42rpl6"
psql -h <pooler-dir>/pg_sockets -p <pg-port> -U postgres -d postgresUser: "connect to multigateway" or "psql multigateway"
psql -h localhost -p <pg-port> -U postgres -d postgres