Skip to main content

Gateway API Implementation Performance Benchmark Plan

📅 Created: 2026-02-12 | Updated: 2026-02-14 | ⏱️ Reading time: ~5 min

A systematic benchmark plan for objectively comparing 5 Gateway API implementations in the same Amazon EKS environment. The goal is to quantitatively identify each solution's strengths and weaknesses to enable data-driven architecture decisions.

Related Document

This benchmark plan targets the 5 solutions compared in the Gateway API Adoption Guide.

1. Benchmark Objective

This benchmark aims to objectively compare 5 Gateway API implementations in the same EKS environment, quantitatively identifying each solution's strengths and weaknesses.

Key Questions:

  • Which solution is the fastest? (throughput, latency)
  • Which solution has the best resource efficiency? (performance relative to CPU/Memory)
  • Which solution scales best in large-scale environments?
  • What are the trade-offs for each solution?

2. Test Environment Design

3. Test Scenarios

1. Basic Throughput (Throughput Test)

Purpose: Measure maximum RPS (Requests Per Second)

Measures maximum throughput for each solution by increasing concurrent connections from 100, 500, 1000, to 5000.

2. Latency Profile

Purpose: Measure P50/P90/P99/P99.9 latency

Measures response time distribution under steady load to compare tail latencies.

3. TLS Performance

Purpose: Measure TLS termination throughput and handshake time

Measures TLS termination performance and handshake overhead for HTTPS traffic.

4. L7 Routing Complexity

Purpose: Measure performance impact of header-based routing and URL rewrite

Measures the impact of complex routing rules on performance.

5. Scaling Test

Purpose: Measure performance changes as route count increases (10, 50, 100, 500 routes)

Measures routing performance and memory usage with many HTTPRoutes.

6. Resource Efficiency

Purpose: Throughput relative to CPU/Memory usage

Compares efficiency of each solution under the same resource constraints.

7. Failure Recovery

Purpose: Traffic impact during controller restart

Measures downtime and recovery time when a Gateway controller restarts.

8. gRPC Performance

Purpose: gRPC streaming throughput

Measures gRPC protocol support and performance.

4. Measured Metrics

MetricUnitMeasurement Method
RPS (Requests Per Second)req/sk6 summary or Prometheus rate()
Latency (P50/P90/P99)msk6 histogram_quantile or Grafana
Error Rate%(failed requests / total requests) x 100
CPU Usage%Prometheus container_cpu_usage_seconds_total
Memory UsageMBPrometheus container_memory_working_set_bytes
Connection Setup Timemsk6 http_req_connecting
TLS Handshake Timemsk6 http_req_tls_handshaking
Network ThroughputMbpsPrometheus rate(container_network_transmit_bytes_total)

5. Expected Results (Theoretical Analysis)

Expected strengths/weaknesses per solution:

AWS Native (ALB + NLB)

  • Strengths: Fully managed, auto-scaling, AWS integration
  • Weaknesses: Latency increase from ALB hop, cost
  • Expected Performance: Medium (throughput 10K RPS, P99 50ms)

Cilium Gateway API (ENI mode)

  • Strengths: Best eBPF performance, native routing, Hubble visibility
  • Weaknesses: Configuration complexity, learning curve
  • Expected Performance: Highest (throughput 30K RPS, P99 15ms)

NGINX Gateway Fabric

  • Strengths: Proven NGINX engine, stability, rich features
  • Weaknesses: Higher memory usage
  • Expected Performance: High (throughput 20K RPS, P99 25ms)

Envoy Gateway

  • Strengths: Rich L7 features, extensibility, observability
  • Weaknesses: Resource overhead
  • Expected Performance: Medium-high (throughput 15K RPS, P99 30ms)

kGateway (Solo.io)

  • Strengths: AI routing, enterprise features
  • Weaknesses: Enterprise license required
  • Expected Performance: Medium-high (throughput 18K RPS, P99 28ms)

6. Benchmark Execution Plan

PhaseDescriptionToolsDuration
1. Environment SetupDeploy EKS cluster and 5 solutions separatelyeksctl, Helm2 days
2. Basic TestsMeasure throughput, latencyk6, Prometheus1 day
3. TLS TestsMeasure HTTPS performancek6 (TLS)0.5 day
4. L7 TestsTest complex routing rulesk6 (custom)0.5 day
5. Scale TestsTest route count increasekubectl, k61 day
6. Resource MeasurementCPU/Memory profilingPrometheus, Grafana1 day
7. Result AnalysisData analysis and report writingJupyter, Matplotlib2 days
info

Benchmark execution results will be updated in this document. For related network benchmarks, see CNI Performance Comparison.