Paginating Results with LIMIT and OFFSET in MySQL
Jul 05, 2025 am 02:41 AMMySQL paging is commonly implemented by LIMIT and OFFSET, but its performance is poor under large data volumes. 1. LIMIT controls the number of each page, OFFSET controls the starting position, and the syntax is LIMIT N OFFSET M; 2. Performance problems are caused by excessive records and discarding OFFSET, resulting in low efficiency; 3. Optimization suggestions include using cursor paging, index acceleration, and lazy loading; 4. Cursor paging locates the starting point of the next page through the unique value of the last record of the previous page, avoiding OFFSET, which is suitable for "next page" operations, and is not suitable for random jumps.
Paging is a common requirement when MySQL querying large amounts of data. For example, when the website displays user lists, article lists and other scenarios, usually only a fixed amount of data is displayed on a page. At this time, LIMIT
and OFFSET
need to be used to achieve pagination.

Basic usage: LIMIT controls the number, OFFSET controls the starting position
MySQL's LIMIT
is used to limit the number of records returned by the query, and OFFSET
is used to specify which record to return from. The basic syntax is as follows:

SELECT * FROM table_name ORDER BY id LIMIT N OFFSET M;
-
N
is the number of records to be displayed per page; -
M
is to skip the previous records, that is(當(dāng)前頁碼- 1) * N
For example, if you want to obtain the data on page 3 and display 10 pieces per page, you need to write it as:
SELECT * FROM users ORDER BY id LIMIT 10 OFFSET 20;
Note: If there are no appropriate sorting fields in the table, duplicate or missing data may occur between different pages.
Pagination performance problem: The bigger the OFFSET, the slower it is?
When the data volume is very large, using OFFSET
may cause performance degradation. Because MySQL actually scans the OFFSET LIMIT
bar record and then discards the previous OFFSET
bar.
For example:
SELECT * FROM orders ORDER BY created_at DESC LIMIT 10 OFFSET 100000;
Although this statement seems to only take 10 pieces of data, the database actually needs to find out 100,010 pieces first and then throw away the first 100,000 pieces, which is very inefficient.
Optimization suggestions:
- If the pagination is very deep (such as more than 10,000 pages), you can consider using "cursor-based pagination" instead of
LIMIT/OFFSET
. - Use indexes to speed up sorting and searches, especially fields that are often used to make paging conditions.
- For scenarios that do not require precise jumps, they can be implemented in a lighter way, such as lazy loading.
Introduction to cursor paging: an efficient solution to replace OFFSET
The so-called "cursor paging" refers to positioning the starting point of the next page by a unique value (such as self-increment ID or timestamp) of the last record of the previous page, thereby avoiding the use of OFFSET
.
For example, if id
of the last record on the previous page is 12345, then the next page can be written like this:
SELECT * FROM users WHERE id > 12345 ORDER BY id LIMIT 10;
This method does not slow down as the page number becomes deeper, because it is always checked from a specific value.
Applicable premise:
- There must be an ordered and unique field (such as an autoincrement primary key or timestamp).
- It is not suitable for random jumps (such as jumping directly to page 100), but it is more suitable for operations like "next page".
In general, it is simple and common to implement paging with LIMIT
and OFFSET
, but you should pay attention to performance issues in large data scenarios. If the page is deep or the performance requirements are high, you can consider replacing it with cursor paging and other methods.
The above is the detailed content of Paginating Results with LIMIT and OFFSET in MySQL. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undress AI Tool
Undress images for free

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics

To reset the root password of MySQL, please follow the following steps: 1. Stop the MySQL server, use sudosystemctlstopmysql or sudosystemctlstopmysqld; 2. Start MySQL in --skip-grant-tables mode, execute sudomysqld-skip-grant-tables&; 3. Log in to MySQL and execute the corresponding SQL command to modify the password according to the version, such as FLUSHPRIVILEGES;ALTERUSER'root'@'localhost'IDENTIFIEDBY'your_new

mysqldump is a common tool for performing logical backups of MySQL databases. It generates SQL files containing CREATE and INSERT statements to rebuild the database. 1. It does not back up the original file, but converts the database structure and content into portable SQL commands; 2. It is suitable for small databases or selective recovery, and is not suitable for fast recovery of TB-level data; 3. Common options include --single-transaction, --databases, --all-databases, --routines, etc.; 4. Use mysql command to import during recovery, and can turn off foreign key checks to improve speed; 5. It is recommended to test backup regularly, use compression, and automatic adjustment.

When handling NULL values ??in MySQL, please note: 1. When designing the table, the key fields are set to NOTNULL, and optional fields are allowed NULL; 2. ISNULL or ISNOTNULL must be used with = or !=; 3. IFNULL or COALESCE functions can be used to replace the display default values; 4. Be cautious when using NULL values ??directly when inserting or updating, and pay attention to the data source and ORM framework processing methods. NULL represents an unknown value and does not equal any value, including itself. Therefore, be careful when querying, counting, and connecting tables to avoid missing data or logical errors. Rational use of functions and constraints can effectively reduce interference caused by NULL.

Turn on MySQL slow query logs and analyze locationable performance issues. 1. Edit the configuration file or dynamically set slow_query_log and long_query_time; 2. The log contains key fields such as Query_time, Lock_time, Rows_examined to assist in judging efficiency bottlenecks; 3. Use mysqldumpslow or pt-query-digest tools to efficiently analyze logs; 4. Optimization suggestions include adding indexes, avoiding SELECT*, splitting complex queries, etc. For example, adding an index to user_id can significantly reduce the number of scanned rows and improve query efficiency.

TosecurelyconnecttoaremoteMySQLserver,useSSHtunneling,configureMySQLforremoteaccess,setfirewallrules,andconsiderSSLencryption.First,establishanSSHtunnelwithssh-L3307:localhost:3306user@remote-server-Nandconnectviamysql-h127.0.0.1-P3307.Second,editMyS

GROUPBY is used to group data by field and perform aggregation operations, and HAVING is used to filter the results after grouping. For example, using GROUPBYcustomer_id can calculate the total consumption amount of each customer; using HAVING can filter out customers with a total consumption of more than 1,000. The non-aggregated fields after SELECT must appear in GROUPBY, and HAVING can be conditionally filtered using an alias or original expressions. Common techniques include counting the number of each group, grouping multiple fields, and filtering with multiple conditions.

MySQL transactions and lock mechanisms are key to concurrent control and performance tuning. 1. When using transactions, be sure to explicitly turn on and keep the transactions short to avoid resource occupation and undolog bloating due to long transactions; 2. Locking operations include shared locks and exclusive locks, SELECT...FORUPDATE plus X locks, SELECT...LOCKINSHAREMODE plus S locks, write operations automatically locks, and indexes should be used to reduce the lock granularity; 3. The isolation level is repetitively readable by default, suitable for most scenarios, and modifications should be cautious; 4. Deadlock inspection can analyze the details of the latest deadlock through the SHOWENGINEINNODBSTATUS command, and the optimization methods include unified execution order, increase indexes, and introduce queue systems.

MySQL paging is commonly implemented using LIMIT and OFFSET, but its performance is poor under large data volume. 1. LIMIT controls the number of each page, OFFSET controls the starting position, and the syntax is LIMITNOFFSETM; 2. Performance problems are caused by excessive records and discarding OFFSET scans, resulting in low efficiency; 3. Optimization suggestions include using cursor paging, index acceleration, and lazy loading; 4. Cursor paging locates the starting point of the next page through the unique value of the last record of the previous page, avoiding OFFSET, which is suitable for "next page" operation, and is not suitable for random jumps.
