dkduckkit.dev

Glossary

Short, citation-friendly definitions tied to calculators on duckkit.dev. Each term links to related tools and cross-references.

Latency and SRE

  • Latency budget

    Latency budget: maximum end-to-end time for a user path. Stack DNS, TLS, gateways, RPC, and DB hops so the sum stays under your SLA or SLO target.

  • SLA and SLO (service level agreement vs objective)

    SLA vs SLO: SLA is the customer contract; SLO is the internal numeric target and error budget so you miss SLO before breaching the public SLA.

  • p50 latency (median)

    p50 latency is the median request duration—half faster, half slower. Pair it with p99 because a good median still hides painful tail outliers.

  • p99 latency

    p99 latency: 99th percentile request duration. SREs use tail percentiles for SLOs—averages hide outliers users still hit in production traffic.

  • Headroom (latency)

    Latency headroom: gap between measured path latency and your SLO ceiling. Capacity and rate limits preserve headroom by preventing queue blowups.

  • Cold start (serverless)

    Cold start (serverless): extra latency while FaaS provisions or warms a runtime. Model it as a hop when traffic is sparse or scales from zero.

  • Cross-AZ latency

    Cross-AZ latency: inter-zone networking inside one cloud region. Replicated systems like Kafka add RTT when leaders and followers span zones.

Rate limiting and resilience

  • Fixed window rate limiting

    Fixed window rate limiting counts requests in fixed time buckets. Simple, but aligned clients can spike to roughly double the nominal limit at edges.

  • Sliding window rate limiting

    Sliding window rate limiting rolls request counts over time to reduce fixed-window boundary spikes—more state, usually fairer client enforcement.

  • Leaky bucket

    Leaky bucket shaping outputs traffic at a steady drip, queueing or dropping bursts—contrast with token buckets that accrue idle burst credit.

  • Token bucket

    Token bucket rate limiting: steady refill plus burst capacity. Common at API gateways; pairs with sliding and fixed windows for fair client throttling.

  • Jitter (retries)

    Retry jitter randomizes backoff so many clients do not retry in lockstep. Use it with exponential backoff to dampen thundering herds during incidents.

  • Exponential backoff

    Exponential backoff grows retry delays multiplicatively to shed load during outages. Add jitter so retries from many clients do not line up in time.

  • Retry storm

    Retry storm: synchronized retries multiply traffic until backends fail again. Mitigate with jitter, exponential backoff, budgets, and edge rate limits.

  • HTTP 429 Too Many Requests

    HTTP 429 Too Many Requests means the server is rate limiting you. Slow down, honor Retry-After when present, and respect RateLimit response headers.

  • RateLimit headers (IETF)

    IETF RateLimit header draft standardizes quota signaling for HTTP clients. Many production APIs still expose legacy X-RateLimit-* header families.

Apache Kafka

  • Kafka partition

    Kafka partition: ordered log shard inside a topic. Consumer group parallelism is capped by partition count—central to lag math and cluster sizing.

  • Consumer lag (Kafka)

    Kafka consumer lag measures how far committed offsets trail the log end per partition. Rising lag means consumers cannot keep up or are stuck.

  • Consumer group (Kafka)

    Kafka consumer group: cooperating consumers share partitions—at most one active consumer per partition so per-partition ordering stays intact.

  • Consumer offset (Kafka)

    Kafka consumer offset is the next record position in a partition. Commits define restart points and feed lag as the gap to the current log end.

  • Effective consumers (Kafka lag)

    Effective Kafka consumers equal min(running consumers, partitions). Extra consumer instances stay idle until you add partitions for more parallelism.

  • Consumer poll loop (Kafka)

    Kafka consumer poll loop drives fetch and heartbeats via poll(). If processing exceeds max.poll.interval.ms, the group may revoke your partitions.

  • Consumer rebalance (Kafka)

    Kafka consumer rebalance moves partition assignments when members join or leave. It pauses consumption briefly and frequent churn can grow consumer lag.

  • Replication factor (Kafka)

    Kafka replication factor is how many brokers store each partition replica. It enables failover and multiplies storage plus replication network cost.

  • Log retention (Kafka)

    Kafka log retention sets how long or how large topic data may grow. It bounds disk use and how far slow consumers can lag before data expires.

  • Producer compression (Kafka)

    Kafka producer compression trades CPU for smaller batches on disk and wire. Pick a codec and ratio that match your payload and broker capacity.

  • Throughput vs bandwidth (Kafka sizing)

    Throughput vs bandwidth (Kafka): throughput is achieved bytes per second; bandwidth is link or disk capacity—size brokers above sustained load.

API versioning

  • Semantic Versioning (SemVer)

    Semantic Versioning (SemVer) uses MAJOR.MINOR.PATCH to signal API compatibility: breaking, backward compatible features, and non-breaking fixes.

  • Major, minor, and patch bumps (API)

    API version bumps under SemVer: patch for fixes, minor for compatible features, major for breaks—communicate consumer impact in release notes.

  • Breaking change (API)

    API breaking change: clients fail or misbehave without updates—field removal, stricter validation, or error semantics usually need a major bump.

  • Calendar Versioning (CalVer)

    Calendar Versioning (CalVer) ties versions to release dates. The calendar alone does not imply SemVer compatibility—document breaking changes explicitly.