← 목록으로
Load balancer distributing traffic across multiple servers in a diagram
Load Balancers

Load Balancers Demystified: How They Keep Networks Running

Load balancers distribute traffic across servers to prevent overload and ensure uptime. Here's how they work and when your network needs one.

By Alex Morgan · Updated 2025년 5월 14일

When a website or application receives more traffic than a single server can handle, a load balancer distributes that traffic across multiple servers. It's one of the most fundamental pieces of networking infrastructure for any application that needs to stay available under heavy load — and understanding how load balancers work is essential for anyone building or managing network services.

How Load Balancing Works

A load balancer sits between clients and a pool of backend servers. When a client sends a request, the load balancer decides which server should handle it based on a configured algorithm. The client never communicates directly with backend servers — the load balancer acts as an intermediary, providing a single point of contact while distributing work behind the scenes.

Common Load Balancing Algorithms

  1. Round Robin — Requests are distributed to servers in sequence: Server A, Server B, Server C, Server A, and so on. Simple and effective when servers have equal capacity
  2. Least Connections — New requests go to the server currently handling the fewest active connections. Better for varied workloads
  3. Weighted Round Robin — Like round robin but assigns more requests to more powerful servers based on configured weights
  4. IP Hash — Uses the client's IP address to determine which server receives the request, ensuring the same client always reaches the same server
  5. Least Response Time — Routes to the server with the lowest current response time, dynamically adapting to server performance

Layer 4 vs Layer 7

Load balancers operate at different layers of the networking stack. Layer 4 (transport) load balancers route based on IP address and TCP/UDP port information without inspecting the actual content. They're fast and efficient but can't make routing decisions based on application data.

Layer 7 (application) load balancers inspect the content of requests — HTTP headers, URLs, cookies — and make intelligent routing decisions. They can send API requests to one server pool and static content requests to another, or route based on user authentication status. Most modern load balancers operate at Layer 7.

Health Checks

A load balancer that sends traffic to a crashed server defeats its purpose. Health checks continuously monitor backend servers by sending test requests at regular intervals. If a server fails to respond correctly, the load balancer automatically removes it from the pool and redistributes its traffic. When the server recovers, it's added back. This automatic failover is often the most valuable feature of a load balancer.

Options for US Businesses

Cloud-based load balancers from AWS (ALB/NLB), Google Cloud, and Azure are the simplest to deploy for cloud-hosted applications. For on-premises deployments, F5 BIG-IP and Citrix ADC are enterprise standards, while HAProxy and NGINX provide powerful open-source alternatives that handle millions of connections.