


Abstract: This article discusses the load balancing technology and load distribution strategy used by distributed network servers, and implements a load balancing gateway on FreeBSD based on network address translation, which is applied to our Internet network server to distribute the load to multiple servers. Sharing to solve the high load problem of CPU or I/O caused by a large number of concurrent accesses faced by Internet servers. In order to achieve the best load balancing effect, the load controller needs to allocate load according to the current CPU and I/O status of each server. This requires dynamically monitoring the load of the server and applying optimized load distribution strategies to achieve even load distribution. Purpose.
Keywords: Load balancing, network address translation, FreeBSD
1. Introduction
The rapid growth of the Internet has caused a rapid increase in the number of visits faced by multimedia network servers , the server needs to have the ability to provide a large number of concurrent access services, and the server's processing and I/O capabilities have become the bottleneck in providing services. Since the performance of a single server is always limited, multi-server and load balancing technology must be used to meet the needs of a large number of concurrent accesses.
The earliest load balancing technology is implemented through DNS. The same name is configured for multiple addresses in DNS, so the client querying this name will get one of the addresses, thus allowing different customers to access different servers to achieve load balancing [1]. DNS load balancing is a simple and effective method, but it cannot distinguish the differences between servers, nor can it reflect the current operating status of the server.
The reverse proxy server can forward requests to internal web servers. If the proxy server can forward requests evenly to multiple internal servers, load balancing can be achieved [2]. In the reverse proxy mode, an optimized load balancing strategy can be applied, and the most idle internal server is accessed each time to provide services. However, as the number of concurrent connections increases, the load on the proxy server itself becomes very large, and eventually the reverse proxy server itself will become a bottleneck of the service.
An address translation gateway that supports load balancing can map an external IP address to multiple internal IP addresses, and dynamically use one of the internal addresses for each TCP connection request to achieve load balancing [3]. Many hardware manufacturers integrate this technology into their switches as a function of their fourth layer switching. They generally use a load balancing strategy of random selection and selection based on the number of server connections or response time to distribute the load. However, the load controller implemented by hardware is not very flexible and cannot support more optimized load balancing strategies and more complex application protocols.
In addition to these three load balancing methods, some protocols internally support load balancing-related functions, such as redirection capabilities in the HTTP protocol, etc., but it depends on the specific protocol and therefore has a limited scope of use. Based on these existing load balancing technologies, we chose to use software to implement network address translation load balancing to make up for the inflexibility of the hardware load balancer, and applied optimized balancing strategies to achieve the best load sharing of back-end servers. Excellent status.
2. Load balancing strategy
In order to evenly distribute the load to multiple internal servers, a certain load balancing strategy needs to be applied. Traditional load balancing strategies do not take into account the different types of service requests, the different capabilities of the backend servers, and the uneven load distribution caused by random selection. In order to make the load distribution very even, it is necessary to apply a load balancing strategy that can correctly reflect the CPU and I/O status of each server [4].
There are various types of service requests initiated by customers. According to the resource requirements for processor, network and I/O, they can be simply divided into two different categories in order to apply different processing strategies. :
Static document request: such as ordinary text, images and other static multimedia data, which have little impact on the processor load, and the resulting disk I/O load is proportional to the size of the document. Proportional, mainly putting pressure on network I/O.
Dynamic document requests: More common requests often require pre-processing by the server, such as searching databases, compressing and decompressing multimedia files, etc. These requests require considerable processor and disk I/O resources. .
For static documents, each service process occupies roughly the same system resources, so the number of processes can be used to represent the system load. The dynamic document service requires additional processing, which occupies more system resources than processing static requests, so it needs to be represented by a weight. Such a simplest server load expression formula is:
where L is the load of the server, Ns is the number of static document service processes, Nd is the number of dynamic document service processes, and a is the relative number of each dynamic document service For the weight of static document service, you can choose between 10 and 100.
The limitation of server hardware is not considered in this formula. When the hardware limit is reached, the load of the server will increase significantly due to resource constraints. For example, due to the limitation of server memory size, some processes must be swapped to the hard disk, causing the system load to increase rapidly.Taking into account the system hardware limitations, the server load can be expressed as:
The newly added parameter Ll represents the limit of the normal load of the server, which should be set according to the hardware capabilities of each server itself. And b represents the weight used to limit the allocation of server tasks when the normal load is exceeded. It should be set to a value greater than Ll to represent the effect of hardware limitations. Usually in a server cluster, the weight of the server with worse hardware settings should be set larger to avoid that when all servers are overloaded, the server with the worst hardware will have the highest load. Therefore, b is inversely proportional to the hardware limit Ll of this server, so b can be set as:
Llmax is the Ll value of the server with the highest hardware configuration in the server cluster. After the load of each server is determined, the central server that controls load distribution can correctly distribute the load to the idlest server, thus preventing uneven load distribution like other load distribution strategies.
3. Implementation method and experimental results
Our server system consists of multiple FreeBSD systems connected using Fast Ethernet. Each backend server runs a daemon process to dynamically obtain its own load status, and the central control gateway implemented using FreeBSD refreshes the load of each server through these daemon processes to perform correct load distribution.
3.1 Gateway that supports load balancing
Under the FreeBSD system, the divert interface is provided to support network address translation capabilities. IP data packets are sent to the divert interface through the ipfw filtering function of the system kernel, so that the external daemon natd can receive the original data packets, process them, and then send them back to the system kernel for normal IP distribution [5].
Therefore, according to the address translation structure of FreeBSD, you can create your own network address translation daemon to support the load balancing function, so that the FreeBSD system can be used as a gateway that supports load balancing. Because it is implemented in software, it can easily support non-standard protocols and apply optimized load balancing strategies, and has great flexibility.
3.2 Experiments and Analysis
To test the usability of this implementation, we conducted our test experiments on the most common HTTP protocols. In order to distinguish different request types, three different types of tests were designed to test different aspects of performance.
Dynamic document generated by CGI program: used to test the load balancing status of the server's processing capabilities.
Small static document: Use a smaller static document to test the load balancing status under frequent connections;
Large static document: Use a larger document , test the load balancing status of disk and network I/O;
The test results are based on the performance of a single server completing requests per second, showing the types of load balancing per second when using multiple servers The ratio of completed requests to the baseline number of requests.
Figure 1: Load balancing performance
The first curve a in the above figure is for processing dynamic document requests. As the number of servers increases, its performance increases exponentially; while the second curve b is for processing small-size static document requests, the performance improvement is not obvious when using three servers; while processing large-size static document requests The third curve c has almost no performance change. In order to find the reason why the load balancing system cannot reach the ideal state, we examined the utilization of server resources:
Table 1. Utilization of server resources
Processing type
Load Balancing Gateway
Server 1
Server 2
Server 3
a
53%
97%
95%
98%
b
76%
43%
39%
41%
c
94%
32%
31%
35%
As can be seen from this table, when processing dynamic document a, all three servers are running at full speed and the load is evenly distributed. This is an ideal state. When processing static document types b and c, although the load is evenly distributed among the three servers, each server is not running at full speed. Especially when processing large-size documents, the natd process in the load balancing device occupies most of the processing resources. Since all network traffic must be converted through it, the load on the natd process increases when the number of network traffic and concurrent connections is considerable. When using different numbers of backend servers in the experiment, the actual network bandwidth flowing through the load balancing gateway is:
Table 2: Bandwidth of the server cluster when serving large-size documents
servers Quantity
1 unit
2 units
3 units
Network speed (Kb/s)
10042.14
11015.10
11442.67
It can be seen that the bandwidth is limited to about 10MB/s. Obviously this is the bandwidth limit of the load balancing process used in this test. In fact, the program uses a linked list to maintain the status of network address translation, which greatly limits Its network performance can be further improved by improving hardware performance and improving algorithms.
4. Discussion
As can be seen from the above experiment, the load balancer based on network address translation can effectively solve the CPU and disk I/O load on the server side. However, The performance of the load balancer itself is limited by network I/O and has certain bandwidth limitations under certain hardware conditions. However, this bandwidth limit can be increased by improving the algorithm and improving the performance of the hardware running the load balancing program. It can also be seen that different service types occupy different server resources. The load measurement strategy we use is to use the same load for evaluation, which is suitable for most conditions. However, the best way is to target different Resources, such as CPU, disk I/O or network I/O, monitor server load respectively, and the central controller selects the most appropriate server to distribute customer requests. Our future work will start from these two aspects to improve this load balancing controller.
References:
[1] E.Kata, M.Butler, and R. McGrath. A scalable HTTP server: the ncsa prototype. Computer Networks and ISDN systems. 1994. Vol 27, P155-164
[2] Ralf S.Engelschall. Load Balancing Your Web Site. Web Techniques Magazine (http://www.WebTechniques.com), May 1998. vol. 3, iss.5
[3] CICSO. LocalDirector Documents. http://www.cisco.com, 1997
[4] H.Zhu. T.Yang, Q.Zheng , D.Watson, O.H.Ibarra, andT.Smith, Adaptive load sharing for clustered digital library servers. Technical Report, CS, UCSB, 1998.
[5] FreeBSD core team. natd and divert manual pages. http ://www.freebsd.org. 1995
Implement a load balancing gateway by NAT
Wang, Bo
NongYe Road 70 , ZhengZhou, 450002, P.R.China
wb@email.online.ha.cn
Abstract: This paper investigates load balancing techniques and strategies, and implements a load balancing gateway based NAT for our Internet servers. The Internet servers involve the high load of CPU and I/O by simultaneous access requests, the symmetrical clustered servers can distribute the server load to solve the problem. To balance the load in the best way, the gateway distributes the load according to the status of server's CPU and I/O. The gateway must monitor every server's load and apply the best scheme to deliver every request, so it can provide the high performance for Internet services.
Keywords: load balancing, NAT, FreeBSD

命名管道是一种在操作系统中相对比较低级的进程通信方式,它是一种以文件为中介的进程通信方式。在Go语言中,通过os包提供了对命名管道的支持。在本文中,我们将介绍如何在Go中使用命名管道来实现进程间通信。一、命名管道的概念命名管道是一种特殊的文件,可以被多个进程同时访问。在Linux系统中,命名管道是一种特殊的文件类型,它们存在于文件系统的某个位置上,并且可以在

在Go语言中,使用第三方库是非常方便的。许多优秀的第三方库和框架可以帮助我们快速地开发应用程序,同时也减少了我们自己编写代码的工作量。但是如何正确地使用第三方库,确保其稳定性和可靠性,是我们必须了解的一个问题。本文将从以下几个方面介绍如何使用第三方库,并结合具体例子进行讲解。一、第三方库的获取Go语言中获取第三方库有以下两种方式:1.使用goget命令首先

随着传统的多线程模型在高并发场景下的性能瓶颈,协程成为了PHP编程领域的热门话题。协程是一种轻量级的线程,能够在单线程中实现多任务的并发执行。在PHP的语言生态中,协程得到了广泛的应用,比如Swoole、Workerman等框架就提供了对协程的支持。那么,如何在PHP中使用协程呢?本文将介绍一些基本的使用方法以及常见的注意事项,帮助读者了解协程的运作原理,以

数据聚合函数是一种用于处理数据库表中多行数据的函数。在PHP中使用数据聚合函数可以使得我们方便地进行数据分析和处理,例如求和、平均数、最大值、最小值等。下面将介绍如何在PHP中使用数据聚合函数。一、介绍常用的数据聚合函数COUNT():计算某一列的行数。SUM():计算某一列的总和。AVG():计算某一列的平均值。MAX():取出某一列的最大值。MIN():

变量函数是指可以使用变量来调用函数的一种特殊语法。在PHP中,变量函数是非常有用的,因为它可以让我们更加灵活地使用函数。在本文中,我们将介绍如何在PHP中使用变量函数。定义变量函数在PHP中,变量函数的定义方式非常简单,只需要将要调用的函数名赋值给一个变量即可。例如,下面的代码定义了一个变量函数:$func='var_dump';这里将var_dump函

<p>Windows 系统上的 OneDrive 应用程序允许您将文件存储在高达 5 GB 的云上。OneDrive 应用程序中还有另一个功能,它允许用户选择一个选项,是将文件保留在系统空间上还是在线提供,而不占用您的系统存储空间。此功能称为按需文件。在这篇文章中,我们进一步探索了此功能,并解释了有关如何在 Windows 11 电脑上的 OneDrive 中按需使用文件的各种选项。</p><h2>如何使用 On

随着音频处理在各种应用场景中的普及,越来越多的程序员开始使用Go编写音频处理程序。Go语言作为一种现代化的编程语言,具有优秀的并发性和高效率的特点,使用它进行音频处理十分方便。本文将介绍如何在Go中使用音频处理技术,包括读取、写入、处理和分析音频数据等方面的内容。一、读取音频数据在Go中读取音频数据有多种方式。其中比较常用的是使用第三方库进行读取,比如go-

近年来,WebSocket技术已经成为了Web开发中不可或缺的一部分。WebSocket是一种在单个TCP连接上进行全双工通信的协议,它使得客户端和服务器之间的通信更加流畅和高效。如今,很多现代的Web应用程序都使用了WebSocket技术,例如实时聊天、在线游戏以及实时数据可视化等。Go语言作为一个现代的编程语言,自然也提供了很好的支持WebSock


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

ZendStudio 13.5.1 Mac
Powerful PHP integrated development environment

mPDF
mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

SecLists
SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

WebStorm Mac version
Useful JavaScript development tools

DVWA
Damn Vulnerable Web App (DVWA) is a PHP/MySQL web application that is very vulnerable. Its main goals are to be an aid for security professionals to test their skills and tools in a legal environment, to help web developers better understand the process of securing web applications, and to help teachers/students teach/learn in a classroom environment Web application security. The goal of DVWA is to practice some of the most common web vulnerabilities through a simple and straightforward interface, with varying degrees of difficulty. Please note that this software
