For some systems with large amounts of data, the database faces not only low query efficiency but also long data storage time. Especially for reporting systems, the time spent on data import may last for several hours or more than ten hours every day. Therefore, it makes sense to optimize database insertion performance.
Recommended: "mysql tutorial"
After some performance tests on MySQL InnoDB, I found some methods that can improve insert efficiency for your reference.
1. Insert multiple pieces of data with one SQL statement
Commonly used insert statements such as:
INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('0', 'userid_0', 'content_0', 0);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('1', 'userid_1', 'content_1', 1);<br/>
are modified to:
INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('0', 'userid_0', 'content_0', 0), ('1', 'userid_1', 'content_1', 1);<br/>
The modified insertion operation can improve the insertion efficiency of the program. The main reason why the second SQL execution efficiency is high here is that the amount of logs after merging (MySQL's binlog and innodb's transaction logs) is reduced, which reduces the amount and frequency of log flushing, thereby improving efficiency. By merging SQL statements, it can also reduce the number of SQL statement parsing and reduce network transmission IO.
Here are some test comparison data, which are to import a single piece of data and convert it into a SQL statement for import, and to test 100, 1,000, and 10,000 data records respectively.
#2. Perform insertion processing in the transaction.
Change the insertion to:
START TRANSACTION;
INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('0', 'userid_0', 'content_0', 0);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('1', 'userid_1', 'content_1', 1);<br/>...
COMMIT;
Using transactions can improve the efficiency of data insertion. This is because when an INSERT operation is performed, a transaction will be established internally in MySQL, and only within the transaction Perform actual insert processing operations. By using transactions, you can reduce the cost of creating transactions. All inserts are executed before committing.
A test comparison is also provided here, which is the case of not using transactions and using transactions when the number of records is 100, 1000, and 10000.
#3. Data is inserted in order.
Orderly insertion of data means that the inserted records are arranged in order on the primary key. For example, datetime is the primary key of the record:
INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('1', 'userid_1', 'content_1', 1);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('0', 'userid_0', 'content_0', 0);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('2', 'userid_2', 'content_2',2);<br/>
is modified to:
INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('0', 'userid_0', 'content_0', 0);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`) <br/> VALUES ('1', 'userid_1', 'content_1', 1);<br/>INSERT INTO `insert_table` (`datetime`, `uid`, `content`, `type`)
VALUES ('2', 'userid_2', 'content_2',2);
Since the database needs to maintain index data when inserting, disordered records will increase the cost of maintaining the index. We can refer to the B tree index used by InnoDB. If each inserted record is at the end of the index, the index positioning efficiency is very high, and the index adjustment is small; if the inserted record is in the middle of the index, the B tree needs to be split and merged. Waiting for processing will consume more computing resources, and the index positioning efficiency of inserted records will decrease. When the amount of data is large, there will be frequent disk operations.
The following provides a performance comparison of random data and sequential data, which are recorded as 100, 1000, 10000, 100000 and 1 million respectively.
From the test results, the performance of this optimization method has been improved, but the improvement is not very obvious.
4. Comprehensive performance test
Here provides a test that uses the above three methods at the same time to optimize INSERT efficiency.
It can be seen from the test results that the performance improvement of the method of merging data transactions is obvious when the amount of data is small. When the amount of data is large (10 million Above), the performance will drop sharply. This is because the amount of data exceeds the capacity of innodb_buffer at this time. Each index positioning involves more disk read and write operations, and the performance drops quickly. The method of using merged data transactions to order data still performs well when the data volume reaches tens of millions. When the data volume is large, the positioning of the ordered data index is more convenient and does not require frequent read and write operations on the disk, so it can Maintain high performance.
Notes:
1. SQL statements have a length limit. When merging data, the SQL length limit must not be exceeded in the same SQL. It can be modified through the max_allowed_packet configuration. , the default is 1M, and was modified to 8M during testing.
2. Transactions need to be controlled in size. If a transaction is too large, it may affect execution efficiency. MySQL has the innodb_log_buffer_size configuration item. If this value is exceeded, the innodb data will be flushed to the disk. At this time, the efficiency will decrease. So a better approach is to commit the transaction before the data reaches this value.
The above is the detailed content of Detailed explanation of MySQL batch SQL insert performance optimization. For more information, please follow other related articles on the PHP Chinese website!

本篇文章给大家带来了关于mysql的相关知识,其中主要介绍了关于架构原理的相关内容,MySQL Server架构自顶向下大致可以分网络连接层、服务层、存储引擎层和系统文件层,下面一起来看一下,希望对大家有帮助。

在mysql中,可以利用char()和REPLACE()函数来替换换行符;REPLACE()函数可以用新字符串替换列中的换行符,而换行符可使用“char(13)”来表示,语法为“replace(字段名,char(13),'新字符串') ”。

方法:1、利用right函数,语法为“update 表名 set 指定字段 = right(指定字段, length(指定字段)-1)...”;2、利用substring函数,语法为“select substring(指定字段,2)..”。

mysql的msi与zip版本的区别:1、zip包含的安装程序是一种主动安装,而msi包含的是被installer所用的安装文件以提交请求的方式安装;2、zip是一种数据压缩和文档存储的文件格式,msi是微软格式的安装包。

转换方法:1、利用cast函数,语法“select * from 表名 order by cast(字段名 as SIGNED)”;2、利用“select * from 表名 order by CONVERT(字段名,SIGNED)”语句。

本篇文章给大家带来了关于mysql的相关知识,其中主要介绍了关于MySQL复制技术的相关问题,包括了异步复制、半同步复制等等内容,下面一起来看一下,希望对大家有帮助。

本篇文章给大家带来了关于mysql的相关知识,其中主要介绍了mysql高级篇的一些问题,包括了索引是什么、索引底层实现等等问题,下面一起来看一下,希望对大家有帮助。

在mysql中,可以利用REGEXP运算符判断数据是否是数字类型,语法为“String REGEXP '[^0-9.]'”;该运算符是正则表达式的缩写,若数据字符中含有数字时,返回的结果是true,反之返回的结果是false。


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

SAP NetWeaver Server Adapter for Eclipse
Integrate Eclipse with SAP NetWeaver application server.

EditPlus Chinese cracked version
Small size, syntax highlighting, does not support code prompt function

Dreamweaver Mac version
Visual web development tools

Notepad++7.3.1
Easy-to-use and free code editor

VSCode Windows 64-bit Download
A free and powerful IDE editor launched by Microsoft
