Load Balancing

The distribution of incoming network traffic across multiple servers to ensure no single server is overwhelmed.

Load balancing is the practice of distributing incoming requests across a pool of backend servers to optimize resource utilization, maximize throughput, minimize response time, and avoid overloading any single server. Load balancers act as a reverse proxy, sitting between clients and servers.

Common load balancing algorithms include round-robin (distributing requests evenly in rotation), least connections (sending to the server with fewest active connections), weighted distribution (sending more traffic to more powerful servers), and IP hash (routing the same client to the same server for session affinity).

Load balancers also perform health checks on backend servers and automatically remove unhealthy servers from the pool. This is a form of automatic failover at the application layer. Modern cloud platforms provide managed load balancing services (AWS ALB/NLB, GCP Load Balancing, Cloudflare). Monitoring the load balancer endpoint with Hyperping ensures you detect issues even when individual backend servers might still be healthy.

Hyperping monitoring dashboard

Related Terms

Failover
The automatic switching to a backup system when the primary system fails, ensuring service continuit...
Health Check
An endpoint or process that verifies whether a service or its dependencies are functioning correctly...
Throughput
The rate at which a system processes requests or data, typically measured in requests per second.
Availability
The proportion of time a system is functional and accessible, often expressed as a percentage.
Auto-Scaling
The automatic adjustment of compute resources based on current demand to maintain performance and co...

Related Resources

Get started

Start monitoring in the next 5 minutes.

Stop letting customers discover your outages first. Set up monitoring, status pages, on-call, and alerts before your next coffee break.

14 days free trial — No card required