


How to solve the duplication problem in mysql using left join connection
Mysql uses left join to connect duplicates
Problem description
When using join query, for example, table A is the main table , left connection to table B. What we expect is that the query result will have as many records as there are in table A. However, there may be a result that the total number of records queried is more than the total number of records in table A. , and when the query results are displayed, some columns are repeated. Simply put, a Cartesian product is generated.
Question example
Table A is the user table (user), and the fields are:
ID name userid
1 aaaa 10001
2 bbbb 10002
3 ccccc 10003
The B table is the first type of product table (product), and the fields are:
ID title time userid
1 Title 1 2014-01-01 10002
2 Title 2 2014-01-01 10002
3 Title 3 2014-01-01 10001
4 Title 4 2018-03-20 10002
5 Title 5 2018-03-20 10003
At this time, when we used the following sql to execute, we found that
selecct * from user left join product on user.userid=product.userid;
The execution result was actually higher than the total number of records in the user table Many
Solutions to the problem
In fact, this problem can be seen by a discerning person at a glance. Because the keywords of the left join are not unique in the product table, this part of the non-unique data generates a Cartesian product. , resulting in more execution results than expected.
The solution is to use unique keys to associate and do link queries
When mysql uses left join, the data in the right table has duplicate data
LEFT JOIN keyword will Returns all rows from the left table (table_name1), even if there are no matching rows in the right table (table_name2). At this time, there is no problem if the right table (table_name2) has no data or only one piece of data after being filtered by the on keyword.
What I want to talk about is how to deal with duplicate data (complete duplication in business) in the right table (table_name2).
When the right table (table_name2) is filtered by the on keyword and duplicate data appears, the data found at this time will be, right table data * duplicate data, other conditional data in the right table, and the data we need The number of items is different.
My solution is to first query the right table (table_name2) according to the filter field grouping, filter out the same data, and then treat the result as the right table for association
前面脑补 LEFT JOIN (SELECT MODEL_CODE,MODEL_NAME from tm_model GROUP BY MODEL_CODE) tm on tav.model_code = tm.MODEL_CODE 后面脑补
The above is the detailed content of How to solve the duplication problem in mysql using left join connection. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



The main reasons why you cannot log in to MySQL as root are permission problems, configuration file errors, password inconsistent, socket file problems, or firewall interception. The solution includes: check whether the bind-address parameter in the configuration file is configured correctly. Check whether the root user permissions have been modified or deleted and reset. Verify that the password is accurate, including case and special characters. Check socket file permission settings and paths. Check that the firewall blocks connections to the MySQL server.

When MySQL modifys table structure, metadata locks are usually used, which may cause the table to be locked. To reduce the impact of locks, the following measures can be taken: 1. Keep tables available with online DDL; 2. Perform complex modifications in batches; 3. Operate during small or off-peak periods; 4. Use PT-OSC tools to achieve finer control.

1. Use the correct index to speed up data retrieval by reducing the amount of data scanned select*frommployeeswherelast_name='smith'; if you look up a column of a table multiple times, create an index for that column. If you or your app needs data from multiple columns according to the criteria, create a composite index 2. Avoid select * only those required columns, if you select all unwanted columns, this will only consume more server memory and cause the server to slow down at high load or frequency times For example, your table contains columns such as created_at and updated_at and timestamps, and then avoid selecting * because they do not require inefficient query se

In MySQL database, the relationship between the user and the database is defined by permissions and tables. The user has a username and password to access the database. Permissions are granted through the GRANT command, while the table is created by the CREATE TABLE command. To establish a relationship between a user and a database, you need to create a database, create a user, and then grant permissions.

Data Integration Simplification: AmazonRDSMySQL and Redshift's zero ETL integration Efficient data integration is at the heart of a data-driven organization. Traditional ETL (extract, convert, load) processes are complex and time-consuming, especially when integrating databases (such as AmazonRDSMySQL) with data warehouses (such as Redshift). However, AWS provides zero ETL integration solutions that have completely changed this situation, providing a simplified, near-real-time solution for data migration from RDSMySQL to Redshift. This article will dive into RDSMySQL zero ETL integration with Redshift, explaining how it works and the advantages it brings to data engineers and developers.

MySQL cannot run directly on Android, but it can be implemented indirectly by using the following methods: using the lightweight database SQLite, which is built on the Android system, does not require a separate server, and has a small resource usage, which is very suitable for mobile device applications. Remotely connect to the MySQL server and connect to the MySQL database on the remote server through the network for data reading and writing, but there are disadvantages such as strong network dependencies, security issues and server costs.

MySQL has a free community version and a paid enterprise version. The community version can be used and modified for free, but the support is limited and is suitable for applications with low stability requirements and strong technical capabilities. The Enterprise Edition provides comprehensive commercial support for applications that require a stable, reliable, high-performance database and willing to pay for support. Factors considered when choosing a version include application criticality, budgeting, and technical skills. There is no perfect option, only the most suitable option, and you need to choose carefully according to the specific situation.

MySQL database performance optimization guide In resource-intensive applications, MySQL database plays a crucial role and is responsible for managing massive transactions. However, as the scale of application expands, database performance bottlenecks often become a constraint. This article will explore a series of effective MySQL performance optimization strategies to ensure that your application remains efficient and responsive under high loads. We will combine actual cases to explain in-depth key technologies such as indexing, query optimization, database design and caching. 1. Database architecture design and optimized database architecture is the cornerstone of MySQL performance optimization. Here are some core principles: Selecting the right data type and selecting the smallest data type that meets the needs can not only save storage space, but also improve data processing speed.
