Given that MySQL Replication is dual-thread, it is importatnt to recognize how Replication looks when it is broken. There are four main topics is this area
SQL Thread Dies
The SQL Thread is responsible for
- Getting the Next SQL Statement fromt the Relay Logs
- Executing the SQL Statement
- Rotating Relay Logs by Deleting any Relay Log that had all its SQL Entries Executed
If any SQL error happens, the SQL Thread simply dies and the following is posted to its Slave Status:
- Error Number
- Error Message
- SQL statement that experienced the Error
- Current database
- Master Log File where the SQL Originated
- Master Log Position where the SQL Originated
This gives an opportunity to troubleshoot, skip the error, run the SQL statement by hand, start replication back up. Sometimes it may be a SQL-based error, such as error 1062 (Duplicate Key). Other times, it may be related to the Storage Engine or the OS.
To figure out if an SQL statement will break replication, you should take any DML (INSERT, UPDATE, or DELETE) and make a corresponding SELECT using the WHERE clause of the DML. Then, run that SELECT to see if the data you are about to manipulate really exists or not.
I/O Thread Dies
The I/O Thread is responsible for four(4) things:
- Downloading SQL from the Binary Log Entries of a Master
- Recording SQL into its Local Relay Logs as a FIFO queue
- Acknowledging Communication Failure
- Attempting the Reestablish of I/O Thread
Any network latency may cause the I/O Thread to simply die and retry connection. Once a while under those circumstances, the Slave's viewpoint of the Master's log file and position (as logged in its relay logs) may be out-of-sync with what Master actually recorded in its binary logs.
Other side effects may include corrupt relay log entries
- caused by bad network transmission, which can be corrected by running CHANGE MASTER TO from the last SQL statement from the Master that the Slave executed.
- caused by corrupt binary log entries on the Master which was successfully transmitted to the relay logs, which can be corrected by
RESET MASTER;
on the Master to Zap all binary logs
- setting up replication from the new current binary log
- using pt-table-sync to correct differences
Temporary Table Usage
Troubleshooting this is like playing "pin the tail on donkey". Most developers are unaware of this until it happens and you try to fix it not realizing where the cause of this began. Here is the scenation: If you use CREATE TEMPORARY TABLE
on a Master, it will replicate to the Slave. During the time the table is in use, it will be kept in existence in the SQL Thread. If you issue STOP SLAVE;
, the SQL Thread is voluntarily killed along with all temporary tables the SQL Thread was holding. You do not realize that this has occurred until you issue START SLAVE;
and the SQL Threads dies again because the needed temp table no longer exists.
To fix this, you have perform surgery on the master's binary logs and replication as follows:
- Step 01) Locate the exact log file and position the
CREATE TEMPORARY TABLE
was issued on the Master
- Step 02) Locate the name of the database that the
CREATE TEMPORARY TABLE
was meant for
- Create the table using
CREATE TABLE
instead of CREATE TEMPORARY TABLE
- Step 03) Run
CHANGE MASTER TO
using the file and position from Step 01
- Step 04) Run
START SLAVE;
until Replication catches up or another table's nonexistence (due to CREATE TEMPORARY TABLE
) breaks replication for this same issue
- Step 05) If replication breaks again because of
CREATE TEMPORARY TABLE
on a different table, go back to Step 01
Network Inconsiderations
Once upon a time, there was a tendency for MySQL to say Replication was running when, in fact, it was not. This can happen when the network has intermittency that may delay data transmission of binary logs but not severe enough to timeout the I/O Thread. Since the MySQL process can be inconsiderate by being a little insensitive to the network, I affectionately call this "Network Inconsideration". While the bug report on this is closed, it is good to have multiple ways to check MySQL Replication as to its ability to run, especially the I/O Thread. Using MySQL 5.5, you could adjust the sensitivity of the I/O Thread using the the heartbeat and timeout parameters centered around Semisynchronous Replication.
When you shifted the IP within your application, any DB Connections that were open at the moment were totally unaware of the move. A quick netstat | grep -i mysql
would reveal that on Master1.
You would be better off doing the cutover of the IP as follows:
- On Master1
FLUSH HOSTS;
service mysql stop
- Change the IP in the App
- On all web servers,
service httpd restart
In order to safeguard the app from having to edit it for the sake of assigning a new IP, try using a DB VIP instead.
For example:
- Master1 is 10.1.2.30
- Master2 is 10.1.2.40
- Use 10.1.2.70 as the DBVIP
Here is what you can setup
- On Master1, run
ip addr add 10.1.2.70/24 dev eth1
- Use
10.1.2.70
in your app
Then, when it is time to cutover, do the following
- On Master1
FLUSH HOSTS;
service mysql stop
ip addr del 10.1.2.70/24 dev eth1
- On Master2
ip addr add 10.1.2.70/24 dev eth1
SHOW SLAVE STATUS\G
and make sure all final SQL statements from Master1 executed
- On all web servers,
service httpd restart
That way, a cutover would not involve editing any part of the app. Please notice that I did not mention running STOP SLAVE
anywhere because this would allow any final SQL statements to flow over from Master1 to Master2 once mysql is stopped on Master1.
Best Answer
There is a project called MySQL Proxy that is attempting to handle read/write splitting in the proxy layer, but it is not production-ready. You can read more about the issues on this page, paying attention to the known issues section.
For now, as others have noted, you have to handle the routing through your application.