veza/veza-backend-api/internal/monitoring/ledger_metrics.go
senke 3f326e8266
Some checks failed
Veza CI / Rust (Stream Server) (push) Successful in 4m22s
Security Scan / Secret Scanning (gitleaks) (push) Successful in 1m5s
Veza CI / Frontend (Web) (push) Failing after 17m19s
E2E Playwright / e2e (full) (push) Failing after 20m28s
Veza CI / Backend (Go) (push) Successful in 21m31s
Veza CI / Notify on failure (push) Successful in 4s
fix(ci): unblock CI red — gofmt + e2e webserver reuse + orders.hyperswitch_payment_id (Day 4)
Three pre-existing infra issues surfaced by the Day 1→Day 3 push wave.
Each is independent — bundled here because the goal is "ci.yml + e2e.yml
green" before the v1.0.9 tag, and they're all small.

(1) gofmt — ci.yml golangci-lint v2 step

  Five files were unformatted on main. Pre-existing (untouched by my
  Item G work, but the formatter caught them now):
    - internal/api/router.go
    - internal/core/marketplace/reconcile_hyperswitch_test.go
    - internal/models/user.go
    - internal/monitoring/ledger_metrics.go
    - internal/monitoring/ledger_metrics_test.go
  Pure whitespace via `gofmt -w` — no behavior change.

(2) e2e silent-fail — playwright webServer port collision

  The e2e workflow pre-starts the backend in step 9 ("Build + start
  backend API") so it can fail-fast on a non-ok health check. But
  playwright.config.ts had `reuseExistingServer: !process.env.CI` on
  the backend webServer entry — meaning in CI Playwright tried to
  spawn a SECOND backend on port 18080. The spawn collided with
  EADDRINUSE and Playwright silently exited before printing any test
  output. The artifact upload then warned "No files were found"
  because tests/e2e/playwright-report/ never got written, and the job
  ended in `Failure` for an unrelated reason (the artifact upload
  step's GHESNotSupportedError).

  Fix: backend `reuseExistingServer: true` always — workflow + dev
  both pre-start backend on 18080. Vite stays `!CI` because the
  workflow doesn't pre-start it. Comment in playwright.config.ts
  documents the symptom so the next person debugging gets the
  pointer immediately.

(3) orders.hyperswitch_payment_id missing in fresh DBs — migration 080
    skip-branch + 099 ordering drift

  Migration 080 (`add_payment_fields`) wraps its ALTERs in
  "skip if orders doesn't exist". At authoring time orders existed
  earlier in the migration sequence; that ordering has since shifted
  (orders is now created at 099_z_create_orders.sql, AFTER 080).
  Result: in any freshly-migrated DB (CI, fresh dev, future restore
  drills) migration 080 takes the skip branch and the columns are
  never added — even though the Order model and the marketplace code
  rely on them.

  Symptom: every CI run logs
    pq: column "hyperswitch_payment_id" does not exist
  from the periodic ledger_metrics worker. Order checkout would also
  fail to persist payment_id at write time, breaking reconciliation.

  Fix: append-only migration 987 with idempotent
  `ADD COLUMN IF NOT EXISTS` + a partial index on the reconciliation
  hot path. Production envs that did pick up 080 in the original
  order are no-ops; fresh envs converge to the same end state.
  Rollback in migrations/rollback/.

Verified locally:
  $ cd veza-backend-api && go build ./... && VEZA_SKIP_INTEGRATION=1 \
      go test -short -count=1 ./internal/...
  (all green)

SKIP_TESTS=1: backend-only Go + Playwright config + SQL. Frontend
unit tests irrelevant to this commit.

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-27 12:03:55 +02:00

225 lines
9.2 KiB
Go

package monitoring
import (
"context"
"time"
"github.com/prometheus/client_golang/prometheus"
"github.com/prometheus/client_golang/prometheus/promauto"
"go.uber.org/zap"
"gorm.io/gorm"
)
// Ledger-health metrics (v1.0.7 item F, audit P1.8).
//
// Five gauges expose stuck-state counts so ops dashboards + alert
// rules can spot a money-movement pipeline stall before a customer
// does. Sampled every LedgerSamplerInterval via a 60s ticker that
// runs a cheap indexed SELECT COUNT(*) per gauge. The query cost is
// bounded: each clause filters on `status + created_at`, both
// indexed on the relevant tables.
//
// Paired with reconciler_* counters so the dashboard can tell the
// whole story at a glance: "we have N stuck orders and the
// reconciler has resolved M of them today."
//
// Plus two alert rules in config/alertmanager/ledger.yml:
// - ledger_stuck_orders_pending > 0 for 10m → page
// - ledger_orphan_refund_rows > 0 for 5m → page (bug in two-phase
// commit between DB and PSP — immediate ops attention)
var (
// LedgerStuckOrdersPending is the count of orders sitting in
// `pending` past the staleness threshold (30m by default). Should
// be 0 in steady state; non-zero means webhooks from the PSP
// stopped arriving or our endpoint is rejecting them.
LedgerStuckOrdersPending = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_ledger_stuck_orders_pending",
Help: "Orders in 'pending' status older than the staleness threshold (30m). Non-zero triggers ops alert.",
})
// LedgerStuckRefundsPending is the count of refunds with a PSP id
// but still in `pending` past the threshold. Symptom: the refund
// was accepted by Hyperswitch but our webhook handler never
// received the terminal event.
LedgerStuckRefundsPending = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_ledger_stuck_refunds_pending",
Help: "Refunds with hyperswitch_refund_id set but status still 'pending' older than 30m.",
})
// LedgerFailedTransfersAtMaxRetry is the count of seller_transfers
// that exhausted the retry worker's attempts. Non-zero = ops
// investigation required; the Stripe Connect side is stuck
// somehow.
LedgerFailedTransfersAtMaxRetry = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_ledger_failed_transfers_at_max_retry",
Help: "seller_transfers with status='permanently_failed' — retry worker gave up.",
})
// LedgerOrphanRefundRows is THE alert gauge. Non-zero means a
// Refund row exists in 'pending' with no hyperswitch_refund_id
// (past 5m) — i.e., Phase 1 ran, Phase 2 crashed. This is a
// two-phase-commit bug. Page on > 0 for 5m.
LedgerOrphanRefundRows = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_ledger_orphan_refund_rows",
Help: "Refunds pending with empty hyperswitch_refund_id older than 5m. Non-zero = crash between Phase 1 and Phase 2 of RefundOrder, page ops immediately.",
})
// LedgerReversalPendingTransfers tracks rows waiting for the
// Stripe reversal worker (v1.0.7 item B). Non-zero during a
// reversal is normal and transient; sustained > 0 means the
// reversal worker is stuck or Stripe Connect is down.
LedgerReversalPendingTransfers = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_ledger_reversal_pending_transfers",
Help: "seller_transfers in 'reversal_pending' older than 30m — reversal worker is behind.",
})
// --- Reconciler metrics (v1.0.7 item F, covers item C) ---
// ReconcilerActionsTotal counts actions the reconciliation worker
// has taken, labelled by phase. Lets dashboards show "reconciler
// fixed N stuck orders this week" vs "orphan refunds auto-failed
// this week" without re-parsing logs.
ReconcilerActionsTotal = promauto.NewCounterVec(prometheus.CounterOpts{
Name: "veza_reconciler_actions_total",
Help: "Reconciler actions taken, by phase (stuck_orders|stuck_refunds|orphan_refunds).",
}, []string{"phase"})
// ReconcilerOrphanRefundsTotal is the load-bearing counter for
// detecting two-phase-commit bugs. Each orphan refund the
// reconciler auto-fails increments this; a sustained non-zero
// rate is the ledger-health equivalent of a fire alarm.
ReconcilerOrphanRefundsTotal = promauto.NewCounter(prometheus.CounterOpts{
Name: "veza_reconciler_orphan_refunds_total",
Help: "Orphan refunds (Phase 2 crash) auto-failed by the reconciler. Non-zero rate = investigate root cause.",
})
// ReconcilerSweepDurationSeconds measures one RunOnce tick so
// slow sweeps show up in alerting.
ReconcilerSweepDurationSeconds = promauto.NewHistogram(prometheus.HistogramOpts{
Name: "veza_reconciler_sweep_duration_seconds",
Help: "Duration of one ReconcileHyperswitchWorker.RunOnce tick in seconds.",
Buckets: prometheus.ExponentialBuckets(0.1, 2, 10), // 0.1s to ~100s
})
// ReconcilerLastRunTimestamp is set to time.Now().Unix() at the
// end of every RunOnce. Alert rule: `now() - timestamp >
// 2 * RECONCILE_INTERVAL` → worker is dead.
ReconcilerLastRunTimestamp = promauto.NewGauge(prometheus.GaugeOpts{
Name: "veza_reconciler_last_run_timestamp",
Help: "Unix timestamp of the last successful ReconcileHyperswitchWorker tick. Stale = worker dead.",
})
)
// Convenience recorders for the reconciler worker. Called from
// internal/core/marketplace/reconcile_hyperswitch.go so that package
// doesn't import Prometheus directly (keeps marketplace clean of
// observability plumbing).
func RecordReconcilerAction(phase string) {
ReconcilerActionsTotal.WithLabelValues(phase).Inc()
}
func RecordReconcilerOrphanRefund() {
ReconcilerOrphanRefundsTotal.Inc()
}
func RecordReconcilerSweepDuration(d time.Duration) {
ReconcilerSweepDurationSeconds.Observe(d.Seconds())
ReconcilerLastRunTimestamp.Set(float64(time.Now().Unix()))
}
// --- Sampler ---
// LedgerSamplerInterval is how often the sampler re-queries the DB.
// 60s is the sweet spot for our volumes — scrape cost negligible, and
// stale-by-up-to-a-minute is fine for the alert-rule windows (10m /
// 5m).
const LedgerSamplerInterval = 60 * time.Second
// Staleness thresholds match the reconciler's defaults. If ops tunes
// the reconciler thresholds via env vars, the sampler still reports
// against these constants — intentional: the two serve different
// audiences (reconciler = auto-recovery, sampler = human visibility).
// A mismatch means alerts fire while the reconciler has already
// started working on the issue, which is the correct behavior.
const (
ledgerStuckOrderAgeThreshold = 30 * time.Minute
ledgerStuckRefundAgeThreshold = 30 * time.Minute
ledgerOrphanRefundAgeThreshold = 5 * time.Minute
ledgerReversalPendingThreshold = 30 * time.Minute
)
// SampleLedgerHealth runs the five count queries and updates the
// gauges. Safe to call concurrently (gauge writes are atomic). Any
// query error sets the corresponding gauge to -1 — a distinctive
// value that dashboards can filter on ("sampler is broken, don't
// trust the number").
//
// Exposed as a function rather than a method so tests can drive it
// directly against a sqlite in-memory DB.
func SampleLedgerHealth(ctx context.Context, db *gorm.DB, logger *zap.Logger) {
now := time.Now()
sample := func(name string, gauge prometheus.Gauge, query string, args ...interface{}) {
var count int64
if err := db.WithContext(ctx).Raw(query, args...).Scan(&count).Error; err != nil {
logger.Error("ledger sampler: query failed",
zap.String("gauge", name),
zap.Error(err))
gauge.Set(-1)
return
}
gauge.Set(float64(count))
}
sample("stuck_orders_pending", LedgerStuckOrdersPending,
`SELECT COUNT(*) FROM orders
WHERE status = 'pending'
AND hyperswitch_payment_id IS NOT NULL AND hyperswitch_payment_id <> ''
AND created_at < ?`,
now.Add(-ledgerStuckOrderAgeThreshold))
sample("stuck_refunds_pending", LedgerStuckRefundsPending,
`SELECT COUNT(*) FROM refunds
WHERE status = 'pending'
AND hyperswitch_refund_id IS NOT NULL AND hyperswitch_refund_id <> ''
AND created_at < ?`,
now.Add(-ledgerStuckRefundAgeThreshold))
sample("failed_transfers_at_max_retry", LedgerFailedTransfersAtMaxRetry,
`SELECT COUNT(*) FROM seller_transfers WHERE status = 'permanently_failed'`)
sample("orphan_refund_rows", LedgerOrphanRefundRows,
`SELECT COUNT(*) FROM refunds
WHERE status = 'pending'
AND (hyperswitch_refund_id IS NULL OR hyperswitch_refund_id = '')
AND created_at < ?`,
now.Add(-ledgerOrphanRefundAgeThreshold))
sample("reversal_pending_transfers", LedgerReversalPendingTransfers,
`SELECT COUNT(*) FROM seller_transfers
WHERE status = 'reversal_pending'
AND updated_at < ?`,
now.Add(-ledgerReversalPendingThreshold))
}
// ScheduleLedgerHealthSampler runs SampleLedgerHealth once at startup
// (so dashboards aren't blank for the first minute) and then every
// LedgerSamplerInterval until ctx is cancelled.
func ScheduleLedgerHealthSampler(ctx context.Context, db *gorm.DB, logger *zap.Logger) {
ticker := time.NewTicker(LedgerSamplerInterval)
go func() {
defer ticker.Stop()
SampleLedgerHealth(ctx, db, logger)
for {
select {
case <-ctx.Done():
logger.Info("Ledger health sampler stopped")
return
case <-ticker.C:
SampleLedgerHealth(ctx, db, logger)
}
}
}()
logger.Info("Ledger health sampler scheduled",
zap.Duration("interval", LedgerSamplerInterval))
}