Performance tuning of SELECT statements is sometimes a very time-consuming task, and in my opinion it follows the Pareto principle. 20% of the effort will probably give you 80% of the performance improvement, and it may take you 80% of the time to get the other 20% of the performance improvement. Unless you work on Venus, where every day equals 243 days on Earth, there's a good chance that delivery deadlines leave you with insufficient time to tune your SQL queries.
Based on my years of experience writing and running SQL statements, I began developing a checklist that I refer to when trying to improve query performance. I refer to it before doing query planning and reading the documentation for the database I'm using, which can sometimes be complex. My checklist is by no means comprehensive or scientific, it's more of a conservative calculation, but I can say that following these simple steps I do get performance improvements most of the time. Checklist below.
CheckIndex
Indexes should be added to all fields used in the WHERE and JOIN parts of the SQL statement. Take this 3-minute SQL performance test. Regardless of your grade, be sure to read those results with information.
Limit workSize of data set
Check those tables used in the SELECT statement to see if you can apply WHERE Clause to filter. A typical example is a query that performs well when there are only a few thousand rows in the table. But as the application grew, queries slowed down. The solution may be as simple as limiting the query to view data for the current month.
When your query statement has a subquery, pay attention to using filtering on the inner statement of the subquery, not on the outer statement.
Select only the fields you need
Extra fields usually increase the texture of the returned data, resulting in more data being returned to the SQL client. Also:
•When using applications with reporting and analysis capabilities, sometimes reporting performance is low because the reporting tool must aggregate the data it receives in detailed form.
•Occasionally the query may run fast enough, but your problem may be a network-related problem because large amounts of detailed data are sent over the network to the reporting server.
•When using a column-oriented DBMS, only the columns you select are read from disk. The fewer columns you include in your query, the smaller the IO overhead.
Remove unnecessary tables
The reason for removing unnecessary tables is the same as the reason for removing unnecessary fields in the query statement.
Writing SQL statements is a process that usually requires a large number of iterative processes of writing and testing SQL statements. During development, you might add tables to a query, and this might not have any impact on the data returned by the SQL code. Once the SQL is running correctly, I find that many people don't review their scripts and delete tables that have no impact or effect on the final data returned. By removing JOINS operations with unnecessary tables, you reduce a large number of processes that the database must perform. Sometimes, like removing columns, you'll find that the data you reduced comes back through the database.
Removing outer join queries
This is easier said than done, depending on how much impact changing the contents of the table has. One solution is to remove the OUTER JOINS operation by placing placeholders in the rows of both tables. Suppose you have the following tables, which define OUTER JOINS to ensure that all data is returned:
CUSTOMER_NAME | |
---|---|
John Doe | |
Mary Jane | |
Peter Pan | |
Joe Soap |
CUSTOMER_ID | SALES_PERSON |
---|---|
Newbee Smith | |
Oldie Jones | |
Another Oldie | |
Greenhorn |
CUSTOMER_NAME | |
---|---|
NO CUSTOMER | |
John Doe | |
Mary Jane | |
Peter Pan | |
2 | |
1 | |
##0 | Greenhorn |
The above is the detailed content of 6 Simple SQL Optimized SELECT Statements. For more information, please follow other related articles on the PHP Chinese website!

InnoDB uses redologs and undologs to ensure data consistency and reliability. 1.redologs record data page modification to ensure crash recovery and transaction persistence. 2.undologs records the original data value and supports transaction rollback and MVCC.

Key metrics for EXPLAIN commands include type, key, rows, and Extra. 1) The type reflects the access type of the query. The higher the value, the higher the efficiency, such as const is better than ALL. 2) The key displays the index used, and NULL indicates no index. 3) rows estimates the number of scanned rows, affecting query performance. 4) Extra provides additional information, such as Usingfilesort prompts that it needs to be optimized.

Usingtemporary indicates that the need to create temporary tables in MySQL queries, which are commonly found in ORDERBY using DISTINCT, GROUPBY, or non-indexed columns. You can avoid the occurrence of indexes and rewrite queries and improve query performance. Specifically, when Usingtemporary appears in EXPLAIN output, it means that MySQL needs to create temporary tables to handle queries. This usually occurs when: 1) deduplication or grouping when using DISTINCT or GROUPBY; 2) sort when ORDERBY contains non-index columns; 3) use complex subquery or join operations. Optimization methods include: 1) ORDERBY and GROUPB

MySQL/InnoDB supports four transaction isolation levels: ReadUncommitted, ReadCommitted, RepeatableRead and Serializable. 1.ReadUncommitted allows reading of uncommitted data, which may cause dirty reading. 2. ReadCommitted avoids dirty reading, but non-repeatable reading may occur. 3.RepeatableRead is the default level, avoiding dirty reading and non-repeatable reading, but phantom reading may occur. 4. Serializable avoids all concurrency problems but reduces concurrency. Choosing the appropriate isolation level requires balancing data consistency and performance requirements.

MySQL is suitable for web applications and content management systems and is popular for its open source, high performance and ease of use. 1) Compared with PostgreSQL, MySQL performs better in simple queries and high concurrent read operations. 2) Compared with Oracle, MySQL is more popular among small and medium-sized enterprises because of its open source and low cost. 3) Compared with Microsoft SQL Server, MySQL is more suitable for cross-platform applications. 4) Unlike MongoDB, MySQL is more suitable for structured data and transaction processing.

MySQL index cardinality has a significant impact on query performance: 1. High cardinality index can more effectively narrow the data range and improve query efficiency; 2. Low cardinality index may lead to full table scanning and reduce query performance; 3. In joint index, high cardinality sequences should be placed in front to optimize query.

The MySQL learning path includes basic knowledge, core concepts, usage examples, and optimization techniques. 1) Understand basic concepts such as tables, rows, columns, and SQL queries. 2) Learn the definition, working principles and advantages of MySQL. 3) Master basic CRUD operations and advanced usage, such as indexes and stored procedures. 4) Familiar with common error debugging and performance optimization suggestions, such as rational use of indexes and optimization queries. Through these steps, you will have a full grasp of the use and optimization of MySQL.

MySQL's real-world applications include basic database design and complex query optimization. 1) Basic usage: used to store and manage user data, such as inserting, querying, updating and deleting user information. 2) Advanced usage: Handle complex business logic, such as order and inventory management of e-commerce platforms. 3) Performance optimization: Improve performance by rationally using indexes, partition tables and query caches.


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Zend Studio 13.0.1
Powerful PHP integrated development environment

DVWA
Damn Vulnerable Web App (DVWA) is a PHP/MySQL web application that is very vulnerable. Its main goals are to be an aid for security professionals to test their skills and tools in a legal environment, to help web developers better understand the process of securing web applications, and to help teachers/students teach/learn in a classroom environment Web application security. The goal of DVWA is to practice some of the most common web vulnerabilities through a simple and straightforward interface, with varying degrees of difficulty. Please note that this software

EditPlus Chinese cracked version
Small size, syntax highlighting, does not support code prompt function

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Safe Exam Browser
Safe Exam Browser is a secure browser environment for taking online exams securely. This software turns any computer into a secure workstation. It controls access to any utility and prevents students from using unauthorized resources.