Files
vpp-maglev/proto/maglev.proto
Pim van Pelt 4347bb9b05 Bug fixes, config validation, SPA tightening, set-weight UI
This session covers three distinct arcs: correctness bug fixes in the
VPP sync path and frontend reducers, new config validation, and a
large polish pass on the web frontend (tighter layout, backend kebab
dialogs, live grouped-table, live config-reload re-sync).

 - encap for a VIP is now derived from the backend address family,
   not the VIP's. A v6 VIP with v4 backends is programmed as IP6_GRE4
   (not the buggy IP6_GRE6), matching the VPP LB plugin's
   requirement that encap reflects the tunnel inner family. desiredVIP
   gained an Encap field populated in desiredFromFrontend.
 - ActivePoolIndex now requires at least one backend in a pool to be
   BOTH in StateUp AND pb.Weight>0 before the pool counts as active.
   Previously a primary pool with every backend manually zeroed would
   still win over a fallback with weight=100, so fallback traffic
   never materialized. New TestActivePoolIndexWeightedFailover table
   pins the rule in five subcases.
 - SyncLBStateVIP gained a flushAddress parameter threaded through
   reconcileVIP; it forces flush=true on the setASWeight call for a
   specific backend regardless of the usual 0→N heuristic. Wires up
   the explicit [flush] knob the CLI exposes.

 - convertFrontend already enforced that backends within one frontend
   share a family. New cross-frontend pass validateVIPFamilyConsistency
   rejects configs where two frontends share a VIP address but carry
   backends in different families — VPP's LB plugin requires every
   VIP on a prefix to have the same encap type, so such a config
   would fail at lb_add_del_vip_v2 time with VNET_API_ERROR_INVALID
   _ARGUMENT (-73). Catching it at config load turns a silent
   runtime failure into a clear startup error.
 - Two new TestValidationErrors cases pin the behavior: mismatched
   families reject, same-family frontends on one VIP address allowed.

 - Proto adds `bool flush = 5` to SetWeightRequest. The RPC now
   drives a VIP sync immediately after mutating config (fixing the
   latent "weight change only takes effect at the next 30s periodic
   reconcile" gap), passing flushAddress = backend IP when req.Flush
   is true.
 - maglevc grows an optional [flush] token: `set frontend F pool P
   backend B weight N [flush]`. Implementation uses two Run closures
   (runSetFrontendPoolBackendWeight and -Flush) because the tree
   walker only puts slot tokens in args — literal keywords like
   `flush` advance the node but don't appear in the arg list.
 - docs/user-guide.md updated with the [flush] optional and a
   three-paragraph explainer of the graceful-drain vs. flush
   semantics at the VPP level.

 - checker.ListFrontends now sorts alphabetically to match the
   existing sort in ListBackends / ListHealthChecks — RPC responses
   no longer shuffle VIPs per call. cmd/frontend/client.go also
   sorts defensively in refreshAll so an old maglevd build renders
   alphabetically too.
 - backendFromProto was returning out.Transitions[n-1] as the
   LastTransition, but maglevd stores (and the proto carries)
   transitions newest-first, so [n-1] was actually the oldest.
   Reverse on read, which normalizes the client's Transitions slice
   to oldest-first and makes [n-1] genuinely the newest. LastTransition
   now points at the actual latest transition record.
 - applyBackendTransition (Go and TS) derives Enabled = state!="disabled"
   so the two fields stay in lockstep — closed a drift window where
   a recently re-enabled backend still rendered with a stuck
   [disabled] tag. The tag was later removed entirely since state
   and enabled carry the same information.

 - Layout tightened substantially: "FRONTENDS" panel header removed,
   zippy-summary and zippy-body paddings cut, backend-table row
   padding dropped to 2px, per-pool <h3> removed. Pools now live in
   a single consolidated table per frontend with a dedicated "pool"
   column that shows the pool name only on the first row of each
   group — classic grouped-table layout, maximally dense.
 - Description moved inline into the Zippy summary as muted italic
   text, freeing a vertical line per frontend card.
 - formatVIPAddress() helper renders IPv6 VIPs as [addr]:port and
   IPv4 as addr:port, matching RFC 3986 authority syntax.
 - Pools with effective_weight=0 on every backend (standby
   fallbacks, fully-drained primaries) render at opacity 0.35 on
   their non-actions cells; the kebab column stays at full contrast
   because its menu is still fully functional on standby backends.
 - Config-reload propagation: a maglevd config-reload-done log
   event triggers triggerConfigResync() on the frontend side —
   refreshAll() runs off the event-dispatch goroutine, then a
   BrowserEvent{Type:"resync"} is published through the broker.
   writeEvent emits type="resync" as a named SSE frame so the
   SPA's existing addEventListener("resync") handler picks it up
   and calls fetchAllState → replaceAll.
 - recomputeEffectiveWeights in stores/state.ts mirrors the
   server-side health.EffectiveWeights logic so the SPA keeps
   pool.effective_weight correct the moment a backend transitions,
   without waiting for the 30s refresh. Fixed a nasty bug where
   applyBackendEffectiveWeight wrote VIP-scoped vpp-lb-sync-as-*
   event weights into every frontend sharing the backend,
   corrupting frontends with different per-pool configured weights.
   The old log-event reducer was removed; applyConfiguredWeight is
   the narrower replacement used by the kebab set-weight flow.
 - applyBackendTransition calls recomputeEffectiveWeights after
   state updates so pool-failover transitions (primary ⇌ fallback)
   reflect instantly in the UI.

 - Confirmation dialogs via a new Modal primitive
   (Portal-mounted to document.body, escape/click-outside close,
   click-outside debounced on mousedown so mid-row-text-selection
   drags don't dismiss).
 - pause/resume/enable/disable each show a Modal with a consequence
   paragraph explaining what hits live traffic ("will keep existing
   flows", "will flush VPP's flow table", etc.). The disable commit
   button is styled btn-danger red.
 - set-weight action shows a Modal with a range slider (0-100,
   seeded from the current configured weight, accent-colored live
   numeric readout via <output>) plus a flush checkbox and a live-
   swapping note/warn paragraph describing what will happen. On
   commit, the SPA also updates its local store via
   applyConfiguredWeight so the operator sees the new weight
   immediately without waiting for the next refresh.

 - ProbeHeartbeat is now state-aware: ▶ (play) at rest for up/
   down/unknown backends, ⏸ (pause) for paused, ⏹ (stop) for
   disabled/removed, ❤️ (heart) during an in-flight probe.
 - Drop the probe-done event listener — fast probes (<10ms)
   could fire probe-done in the same render tick as probe-start
   and the heart would never visibly paint. Each probe-start now
   runs a fixed 400ms scale-pop animation on a timer; subsequent
   probe-start events reset the timer, so fast cadences produce a
   continuous heart pulse.
 - Fixed wrapper box (16x14 px, overflow hidden) so the row
   doesn't jiggle when the glyph swaps between the narrow ▶/⏸/⏹
   text glyphs and the wider ❤️ emoji.

 - Brand wordmark changed from "maglev" to "vpp-maglev" and wrapped
   in an <a> linking to https://git.ipng.ch/ipng/vpp-maglev. Logo
   link changed to https://ipng.ch/. Both open in a new tab with
   rel="noopener".
 - .gitignore fix: `frontend`, `maglevc`, `maglevd` were matching
   ANY file or directory with those names anywhere in the tree,
   silently ignoring cmd/frontend and friends. Anchored with
   leading slashes so only repo-root build artifacts match.
2026-04-12 23:06:42 +02:00

309 lines
9.4 KiB
Protocol Buffer

syntax = "proto3";
package maglev;
option go_package = "git.ipng.ch/ipng/vpp-maglev/internal/grpcapi";
// Maglev exposes the state of backend health for all frontends.
service Maglev {
rpc ListFrontends(ListFrontendsRequest) returns (ListFrontendsResponse);
rpc GetFrontend(GetFrontendRequest) returns (FrontendInfo);
rpc ListBackends(ListBackendsRequest) returns (ListBackendsResponse);
rpc GetBackend(GetBackendRequest) returns (BackendInfo);
rpc PauseBackend(BackendRequest) returns (BackendInfo);
rpc ResumeBackend(BackendRequest) returns (BackendInfo);
rpc EnableBackend(BackendRequest) returns (BackendInfo);
rpc DisableBackend(BackendRequest) returns (BackendInfo);
rpc ListHealthChecks(ListHealthChecksRequest) returns (ListHealthChecksResponse);
rpc GetHealthCheck(GetHealthCheckRequest) returns (HealthCheckInfo);
rpc SetFrontendPoolBackendWeight(SetWeightRequest) returns (FrontendInfo);
rpc WatchEvents(WatchRequest) returns (stream Event);
rpc CheckConfig(CheckConfigRequest) returns (CheckConfigResponse);
rpc ReloadConfig(ReloadConfigRequest) returns (ReloadConfigResponse);
rpc GetVPPInfo(GetVPPInfoRequest) returns (VPPInfo);
rpc GetVPPLBState(GetVPPLBStateRequest) returns (VPPLBState);
rpc SyncVPPLBState(SyncVPPLBStateRequest) returns (SyncVPPLBStateResponse);
rpc GetVPPLBCounters(GetVPPLBCountersRequest) returns (VPPLBCounters);
}
// ---- requests ---------------------------------------------------------------
message ListFrontendsRequest {}
message GetFrontendRequest {
string name = 1;
}
message ListBackendsRequest {}
message GetBackendRequest {
string name = 1;
}
message BackendRequest {
string name = 1;
}
message ListHealthChecksRequest {}
message GetHealthCheckRequest {
string name = 1;
}
message CheckConfigRequest {}
message CheckConfigResponse {
bool ok = 1;
string parse_error = 2; // set when YAML cannot be read or parsed
string semantic_error = 3; // set when YAML is valid but semantically incorrect
}
message ReloadConfigRequest {}
message ReloadConfigResponse {
bool ok = 1;
string parse_error = 2; // set when YAML cannot be read or parsed
string semantic_error = 3; // set when YAML is valid but semantically incorrect
string reload_error = 4; // set when config is valid but the reload itself failed
}
message GetVPPInfoRequest {}
message VPPInfo {
string version = 1;
string build_date = 2;
string build_directory = 3;
uint32 pid = 4;
int64 boottime_ns = 5; // unix timestamp (ns) when VPP started (from /sys/boottime)
int64 connecttime_ns = 6; // unix timestamp (ns) when maglevd connected to VPP
}
// ---- VPP load-balancer state ------------------------------------------------
message GetVPPLBStateRequest {}
// VPPLBConf mirrors VPP's lb_conf_get_reply: global LB plugin settings.
message VPPLBConf {
string ip4_src_address = 1;
string ip6_src_address = 2;
uint32 sticky_buckets_per_core = 3;
uint32 flow_timeout = 4;
}
// VPPLBAS is one application server attached to a VIP.
message VPPLBAS {
string address = 1;
uint32 weight = 2; // 0-100
uint32 flags = 3; // VPP AS flags (bit 0 = used, bit 1 = flushed)
uint32 num_buckets = 4;
int64 in_use_since_ns = 5; // unix timestamp (ns), 0 if never used
}
// VPPLBVIP mirrors VPP's lb_vip_details plus the attached application servers.
// Note: srv_type, dscp, and target_port are intentionally omitted — maglevd
// only supports GRE encap, so NAT/L3DSR-specific fields don't apply.
message VPPLBVIP {
string prefix = 1; // CIDR, e.g. 192.0.2.1/32
uint32 protocol = 2; // 6=TCP, 17=UDP, 255=any
uint32 port = 3; // 0 = all-port VIP
string encap = 4; // gre4|gre6|l3dsr|nat4|nat6
uint32 flow_table_length = 5;
repeated VPPLBAS application_servers = 6;
bool src_ip_sticky = 7; // source-IP based sticky session (scraped via cli_inband)
}
message VPPLBState {
VPPLBConf conf = 1;
repeated VPPLBVIP vips = 2;
}
// SyncVPPLBStateRequest triggers a reconciliation between the maglev config
// and the VPP load-balancer dataplane. When frontend_name is set, only that
// frontend's VIP is synced (SyncLBStateVIP) and no VIPs are removed. When
// unset, a full reconciliation runs (SyncLBStateAll), which will also remove
// stale VIPs from VPP.
message SyncVPPLBStateRequest {
optional string frontend_name = 1;
}
message SyncVPPLBStateResponse {}
// ---- VPP load-balancer runtime counters ------------------------------------
// GetVPPLBCountersRequest asks maglevd for the most recent per-VIP and
// per-backend counter snapshot. The data is served from an in-process
// cache that is refreshed every ~5 seconds server-side; the call itself
// is cheap and does not hit VPP.
message GetVPPLBCountersRequest {}
// VPPLBVIPCounters is the point-in-time counter row for a single VIP.
// The four lb_* fields are the LB plugin's SimpleCounters (packets only);
// packets / bytes come from the VPP FIB's combined counter at the VIP's
// host prefix (/net/route/to).
message VPPLBVIPCounters {
string prefix = 1; // CIDR, e.g. 192.0.2.1/32
string protocol = 2; // tcp | udp | any
uint32 port = 3;
uint64 next_packet = 4; // "/packet from existing sessions"
uint64 first_packet = 5; // "/first session packet"
uint64 untracked_packet = 6; // "/untracked packet"
uint64 no_server = 7; // "/no server configured"
uint64 packets = 8; // /net/route/to (FIB, summed across workers)
uint64 bytes = 9; // /net/route/to (FIB, summed across workers)
}
// VPPLBBackendCounters is the FIB combined counter for a single backend's
// host prefix, summed across worker threads.
message VPPLBBackendCounters {
string backend = 1; // backend name from config
string address = 2; // backend IP address
uint64 packets = 3;
uint64 bytes = 4;
}
message VPPLBCounters {
repeated VPPLBVIPCounters vips = 1;
repeated VPPLBBackendCounters backends = 2;
}
message SetWeightRequest {
string frontend = 1;
string pool = 2;
string backend = 3;
int32 weight = 4; // 0-100
// flush, when true, also clears VPP's flow table for this backend
// so existing sessions are torn down. When false (default), only
// Maglev's new-bucket mapping is updated and live flows keep
// draining to this backend.
bool flush = 5;
}
// WatchRequest controls which event types are streamed. All fields default to
// true (i.e. an empty request subscribes to everything at info level).
message WatchRequest {
optional bool log = 1; // include log events (default: true)
string log_level = 2; // minimum log level: debug|info|warn|error (default: info)
optional bool backend = 3; // include backend transition events (default: true)
optional bool frontend = 4; // include frontend events (default: true)
}
// ---- responses --------------------------------------------------------------
message ListFrontendsResponse {
repeated string frontend_names = 1;
}
message PoolBackendInfo {
string name = 1;
int32 weight = 2; // configured weight from YAML (0-100)
int32 effective_weight = 3; // state-aware weight after pool-failover logic
}
message PoolInfo {
string name = 1;
repeated PoolBackendInfo backends = 2;
}
message FrontendInfo {
string name = 1;
string address = 2;
string protocol = 3;
uint32 port = 4;
repeated PoolInfo pools = 5;
string description = 6;
bool src_ip_sticky = 7; // VPP LB uses src-IP-based stickiness for this VIP
}
message ListBackendsResponse {
repeated string backend_names = 1;
}
message ListHealthChecksResponse {
repeated string names = 1;
}
message HTTPCheckParams {
string path = 1;
string host = 2;
int32 response_code_min = 3;
int32 response_code_max = 4;
string response_regexp = 5;
string server_name = 6;
bool insecure_skip_verify = 7;
}
message TCPCheckParams {
bool ssl = 1;
string server_name = 2;
bool insecure_skip_verify = 3;
}
message HealthCheckInfo {
string name = 1;
string type = 2;
uint32 port = 3;
string probe_ipv4_src = 4;
string probe_ipv6_src = 5;
int64 interval_ns = 6;
int64 fast_interval_ns = 7;
int64 down_interval_ns = 8;
int64 timeout_ns = 9;
int32 rise = 10;
int32 fall = 11;
HTTPCheckParams http = 12;
TCPCheckParams tcp = 13;
}
message BackendInfo {
string name = 1;
string address = 2;
string state = 3;
repeated TransitionRecord transitions = 4;
bool enabled = 5;
string healthcheck = 6;
}
message TransitionRecord {
string from = 1;
string to = 2;
int64 at_unix_ns = 3;
}
// ---- event stream -----------------------------------------------------------
// LogAttr is a single key/value attribute from a structured log record.
message LogAttr {
string key = 1;
string value = 2;
}
// LogEvent carries a single structured log record.
message LogEvent {
int64 at_unix_ns = 1;
string level = 2;
string msg = 3;
repeated LogAttr attrs = 4;
}
// BackendEvent is emitted on every backend state transition.
message BackendEvent {
string backend_name = 1;
TransitionRecord transition = 2;
}
// FrontendEvent is emitted when a frontend's aggregate state changes.
// Frontends have three states: unknown, up, down. See docs/healthchecks.md.
message FrontendEvent {
string frontend_name = 1;
TransitionRecord transition = 2;
}
// Event is the envelope returned by WatchEvents.
message Event {
oneof event {
LogEvent log = 1;
BackendEvent backend = 2;
FrontendEvent frontend = 3;
}
}