How to Efficiently Select Random Rows from Large PostgreSQL Tables?
PostgreSQL random row selection method
Traditional random row selection methods are inefficient and slow when dealing with large tables containing millions or even billions of records. Two common methods are:
-
Use
random()
to filter:select * from table where random() < 0.001;
Copy after login -
Use
order by random()
andlimit
:select * from table order by random() limit 1000;
Copy after login
However, due to the need for a full table scan or sorting, these methods are not the best choice for tables with a large number of rows and will cause performance bottlenecks.
Optimization methods for large tables
For the following types of tables, consider the following optimization method, which is significantly faster:
- Numeric ID columns with small or medium gaps (indexed for faster lookups)
- No or minimal write operations during selection
Query:
WITH params AS ( SELECT 1 AS min_id, -- 可选:自定义最小ID起始值 5100000 AS id_span -- 近似ID范围(最大ID - 最小ID + 缓冲) ) SELECT * FROM ( SELECT DISTINCT 1 + trunc(random() * p.id_span)::integer AS id FROM params p, generate_series(1, 1100) g GROUP BY 1 ) r INNER JOIN big ON r.id = big.id LIMIT 1000;
How it works:
-
ID range estimate:
- If not known exactly, query the table to estimate the minimum, maximum, and total span (max - min) of the ID column.
-
Random ID generation:
- Generate a different set of random numbers within the estimated ID range.
-
Redundancy and duplication elimination:
- Group the generated numbers to remove duplicates, reducing the possibility of selecting missing rows or already selected rows.
-
Table joins and restrictions:
- Join the random numbers with the actual table using the ID column (must be indexed). This efficient join retrieves the corresponding data for the selected row.
- Finally, apply a limit to retrieve the required number of rows.
Why it’s fast:
-
Minimal index usage:
- The query only performs an index scan on the ID column, which is much faster than a full table scan or sort operation.
-
Optimized random number generation:
- The generated random numbers are distributed over the estimated ID range, minimizing the possibility of missing or overlapping rows.
-
Redundancy and duplication elimination:
- Grouping the generated numbers ensures that only distinct rows are selected, reducing the need for additional filtering or joining to eliminate duplicates.
Other options:
-
Recursive CTE to handle gaps:
- For tables with gaps in the ID sequence, add an additional CTE to handle these gaps.
-
Function wrappers for reuse:
- Define a function that takes limit and gap percentage as parameters, allowing easy configuration and reuse with different tables.
-
Universal functions for any table:
- Create a generic function that accepts any table with integer columns as a parameter.
-
Materialize views for speed:
- Consider creating a materialized view based on an optimized query for faster retrieval of (quasi) randomly selected rows.
-
TABLE SAMPLE
in PostgreSQL 9.5:- Leverage PostgreSQL's "
TABLE SAMPLE SYSTEM
" feature to implement a faster but less random row sampling method, ensuring an accurate number of rows is returned. However, keep in mind that the sample may not be completely random due to clustering effects.
- Leverage PostgreSQL's "
The above is the detailed content of How to Efficiently Select Random Rows from Large PostgreSQL Tables?. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics











Full table scanning may be faster in MySQL than using indexes. Specific cases include: 1) the data volume is small; 2) when the query returns a large amount of data; 3) when the index column is not highly selective; 4) when the complex query. By analyzing query plans, optimizing indexes, avoiding over-index and regularly maintaining tables, you can make the best choices in practical applications.

MySQL is an open source relational database management system. 1) Create database and tables: Use the CREATEDATABASE and CREATETABLE commands. 2) Basic operations: INSERT, UPDATE, DELETE and SELECT. 3) Advanced operations: JOIN, subquery and transaction processing. 4) Debugging skills: Check syntax, data type and permissions. 5) Optimization suggestions: Use indexes, avoid SELECT* and use transactions.

MySQL is suitable for beginners because it is simple to install, powerful and easy to manage data. 1. Simple installation and configuration, suitable for a variety of operating systems. 2. Support basic operations such as creating databases and tables, inserting, querying, updating and deleting data. 3. Provide advanced functions such as JOIN operations and subqueries. 4. Performance can be improved through indexing, query optimization and table partitioning. 5. Support backup, recovery and security measures to ensure data security and consistency.

The main role of MySQL in web applications is to store and manage data. 1.MySQL efficiently processes user information, product catalogs, transaction records and other data. 2. Through SQL query, developers can extract information from the database to generate dynamic content. 3.MySQL works based on the client-server model to ensure acceptable query speed.

MySQL is an open source relational database management system, mainly used to store and retrieve data quickly and reliably. Its working principle includes client requests, query resolution, execution of queries and return results. Examples of usage include creating tables, inserting and querying data, and advanced features such as JOIN operations. Common errors involve SQL syntax, data types, and permissions, and optimization suggestions include the use of indexes, optimized queries, and partitioning of tables.

InnoDB uses redologs and undologs to ensure data consistency and reliability. 1.redologs record data page modification to ensure crash recovery and transaction persistence. 2.undologs records the original data value and supports transaction rollback and MVCC.

MySQL's position in databases and programming is very important. It is an open source relational database management system that is widely used in various application scenarios. 1) MySQL provides efficient data storage, organization and retrieval functions, supporting Web, mobile and enterprise-level systems. 2) It uses a client-server architecture, supports multiple storage engines and index optimization. 3) Basic usages include creating tables and inserting data, and advanced usages involve multi-table JOINs and complex queries. 4) Frequently asked questions such as SQL syntax errors and performance issues can be debugged through the EXPLAIN command and slow query log. 5) Performance optimization methods include rational use of indexes, optimized query and use of caches. Best practices include using transactions and PreparedStatemen

MySQL is chosen for its performance, reliability, ease of use, and community support. 1.MySQL provides efficient data storage and retrieval functions, supporting multiple data types and advanced query operations. 2. Adopt client-server architecture and multiple storage engines to support transaction and query optimization. 3. Easy to use, supports a variety of operating systems and programming languages. 4. Have strong community support and provide rich resources and solutions.
