MySQL
这里使用Percona 版本开启来了参数log_slow_verbosity,得到了更详细的慢查询信息。通常情况下信息没有这么多,但是一定是包含关系,使用Percona的参数解释说明一下这个参数的含义。

一、慢查询中的时间

实际上慢查询中的时间就是时钟时间,是通过操作系统的命令获得的时间,如下是Linux中获取时间的方式

  1. while (gettimeofday(&t, NULL) != 0)
  2. {}
  3. newtime= (ulonglong)t.tv_sec * 1000000 + t.tv_usec;
  4. return newtime;

实际上就是通过OS的API gettimeofday函数获得的时间。

二、慢查询记录的依据

  1. long_query_time:如果执行时间超过本参数设置记录慢查询。
  2. log_queries_not_using_indexes:如果语句未使用索引记录慢查询。
  3. log_slow_admin_statements:是否记录管理语句。(如ALTER TABLE,ANALYZE TABLE, CHECK TABLE, CREATE INDEX, DROP INDEX, OPTIMIZE TABLE, and REPAIR TABLE.)

本文主要讨论long_query_time参数的含义。

三、long_query_time参数的具体含义

如果将语句的执行时间定义为如下:
实际消耗时间 = 实际执行时间+锁等待消耗时间
那么long_query_time实际上界定的是实际执行时间,所以有些情况下虽然语句实际消耗的时间很长但是是因为锁等待时间较长而引起的,那么实际上这种语句也不会记录到慢查询。
看一下log_slow_applicable函数的代码片段:

  1. res= cur_utime - thd->utime_after_lock;
  2. if (res > thd->variables.long_query_time)
  3. thd->server_status|= SERVER_QUERY_WAS_SLOW;
  4. else
  5. thd->server_status&= ~SERVER_QUERY_WAS_SLOW;

这里实际上清楚的说明了上面的观点,是不是慢查询就是通过这个函数进行的判断的,非常重要。可以清晰的看到如下公式:

  • res (实际执行时间 ) = cur_utime(实际消耗时间) - thd->utime_after_lock( 锁等待消耗时间)

实际上在慢查询中记录的正是

  • Query_time:实际消耗时间
  • Lock_time:锁等待消耗时间

但是是否是慢查询其评判标准却是实际执行时间及Query_time - Lock_time
其中锁等待消耗时间( Lock_time)现在已经知道的包括:

  1. MySQL层 MDL LOCK等待消耗的时间。(Waiting for table metadata lock)
  2. MySQL层 MyISAM表锁消耗的时间。(Waiting for table level lock)
  3. InnoDB层 行锁消耗的时间。

    四、MySQL是如何记录锁时间

    可以看到在公式中utime_after_lock( 锁等待消耗时间Lock_time)的记录也就成了整个公式的关键,那么试着进行debug。

    1、MySQL层utime_after_lock的记录方式

    不管是 MDL LOCK等待消耗的时间还是 MyISAM表锁消耗的时间都是在MySQL层记录的,实际上它只是记录在函数mysql_lock_tables的末尾会调用的THD::set_time_after_lock进行的记录时间而已如下:
    1. void set_time_after_lock()
    2. {
    3. utime_after_lock= my_micro_time();
    4. MYSQL_SET_STATEMENT_LOCK_TIME(m_statement_psi, (utime_after_lock - start_utime));
    5. }
    那么这里可以解析为代码运行到mysql_lock_tables函数的末尾之前的所有的时间都记录到utime_after_lock时间中,实际上并不精确。但是实际上MDL LOCK的获取和MyISAM表锁的获取都包含在里面。所以即便是select语句也会看到Lock_time并不为0。下面是栈帧:
    1. #0 THD::set_time_after_lock (this=0x7fff28012820) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_class.h:3414
    2. #1 0x0000000001760d6d in mysql_lock_tables (thd=0x7fff28012820, tables=0x7fff28c16b58, count=1, flags=0) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/lock.cc:366
    3. #2 0x000000000151dc1a in lock_tables (thd=0x7fff28012820, tables=0x7fff28c165b0, count=1, flags=0) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_base.cc:6700
    4. #3 0x00000000017c4234 in Sql_cmd_delete::mysql_delete (this=0x7fff28c16b50, thd=0x7fff28012820, limit=18446744073709551615)
    5. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_delete.cc:136
    6. #4 0x00000000017c84ba in Sql_cmd_delete::execute (this=0x7fff28c16b50, thd=0x7fff28012820) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_delete.cc:1389
    7. #5 0x00000000015a7814 in mysql_execute_command (thd=0x7fff28012820, first_level=true) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:3729
    8. #6 0x00000000015adcd6 in mysql_parse (thd=0x7fff28012820, parser_state=0x7ffff035b600) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:5836
    9. #7 0x00000000015a1b95 in dispatch_command (thd=0x7fff28012820, com_data=0x7ffff035bd70, command=COM_QUERY)
    10. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:1447
    11. #8 0x00000000015a09c6 in do_command (thd=0x7fff28012820) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:1010

    2、InnoDB层的行锁的utime_after_lock记录方式

    InnoDB引擎层调用通过thd_set_lock_wait_time调用thd_storage_lock_wait函数完成的栈帧如下:
    1. #0 thd_storage_lock_wait (thd=0x7fff2c000bc0, value=9503561) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_class.cc:798
    2. #1 0x00000000019a4b2a in thd_set_lock_wait_time (thd=0x7fff2c000bc0, value=9503561)
    3. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:1784
    4. #2 0x0000000001a4b50f in lock_wait_suspend_thread (thr=0x7fff2c088200) at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/lock/lock0wait.cc:363
    5. #3 0x0000000001b0ec9b in row_mysql_handle_errors (new_err=0x7ffff0317d54, trx=0x7ffff2f2e5d0, thr=0x7fff2c088200, savept=0x0)
    6. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/row/row0mysql.cc:772
    7. #4 0x0000000001b4fe61 in row_search_mvcc (buf=0x7fff2c087640 "\377", mode=PAGE_CUR_G, prebuilt=0x7fff2c087ac0, match_mode=0, direction=0)
    8. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/row/row0sel.cc:5940
    9. #5 0x00000000019b3051 in ha_innobase::index_read (this=0x7fff2c087100, buf=0x7fff2c087640 "\377", key_ptr=0x0, key_len=0, find_flag=HA_READ_AFTER_KEY)
    10. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9104
    11. #6 0x00000000019b4374 in ha_innobase::index_first (this=0x7fff2c087100, buf=0x7fff2c087640 "\377")
    12. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9551
    13. #7 0x00000000019b462c in ha_innobase::rnd_next (this=0x7fff2c087100, buf=0x7fff2c087640 "\377")
    14. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9656
    15. #8 0x0000000000f66f1b in handler::ha_rnd_next (this=0x7fff2c087100, buf=0x7fff2c087640 "\377") at /root/mysql5.7.14/percona-server-5.7.14-7/sql/handler.cc:3099
    16. #9 0x00000000014c61b6 in rr_sequential (info=0x7ffff03189e0) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/records.cc:520
    17. #10 0x00000000017c56c3 in Sql_cmd_delete::mysql_delete (this=0x7fff2c006ae8, thd=0x7fff2c000bc0, limit=1)
    18. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_delete.cc:454
    19. #11 0x00000000017c84ba in Sql_cmd_delete::execute (this=0x7fff2c006ae8, thd=0x7fff2c000bc0) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_delete.cc:1389
    函数本身还是很简单自己看看就知道了就是相加而已如下:
    1. void thd_storage_lock_wait(THD *thd, long long value)
    2. {
    3. thd->utime_after_lock+= value;
    4. }

    五、Percona中的log_slow_verbosity参数

    这是Percona的解释:
    Specifies how much information to include in your slow log. The value is a comma-delimited string, and can contain any combination of the following values:
  • microtime: Log queries with microsecond precision (mandatory).
  • query_plan: Log information about the query’s execution plan (optional).
  • innodb: Log InnoDB statistics (optional).
  • minimal: Equivalent to enabling just microtime.
  • standard: Equivalent to enabling microtime,innodb.
  • full: Equivalent to all other values OR’ed together.

总之在Percona中可以修改这个参数获得更加详细的信息大概的格式如下:

  1. # Time: 2018-05-30T09:30:12.039775Z
  2. # User@Host: root[root] @ localhost [] Id: 10
  3. # Schema: test Last_errno: 1317 Killed: 0
  4. # Query_time: 19.254508 Lock_time: 0.001043 Rows_sent: 0 Rows_examined: 0 Rows_affected: 0
  5. # Bytes_sent: 44 Tmp_tables: 0 Tmp_disk_tables: 0 Tmp_table_sizes: 0
  6. # InnoDB_trx_id: 0
  7. # QC_Hit: No Full_scan: No Full_join: No Tmp_table: No Tmp_table_on_disk: No
  8. # Filesort: No Filesort_on_disk: No Merge_passes: 0
  9. # InnoDB_IO_r_ops: 0 InnoDB_IO_r_bytes: 0 InnoDB_IO_r_wait: 0.000000
  10. # InnoDB_rec_lock_wait: 0.000000 InnoDB_queue_wait: 0.000000
  11. # InnoDB_pages_distinct: 0
  12. SET timestamp=1527672612;
  13. select count(*) from z1 limit 1;

六、输出的详细解释

本节将会进行详细的解释,全部的慢查询的输出都来自于函数File_query_log::write_slow ,有兴趣的同学可以自己看看,这里也会给出输出的位置和含义,其中含义部分可能给出的是源码中的注释。

1、第一部分时间

  1. # Time: 2018-05-30T09:30:12.039775Z

对应的代码:

  1. my_snprintf(buff, sizeof buff,"# Time: %s\n", my_timestamp);

其中my_timestamp取值来自于

  1. thd->current_utime();

实际上就是:

  1. while (gettimeofday(&t, NULL) != 0)
  2. {}
  3. newtime= (ulonglong)t.tv_sec * 1000000 + t.tv_usec;
  4. return newtime;

可以看到实际就是调用gettimeofday系统调用得到的系统当前时间。
注意:对于5.6来讲还有一句判断

  1. if (current_time != last_time)

如果两次打印的时间秒钟一致则不会输出时间,只有通过后面介绍的

  1. SET timestamp=1527753496;

来判断时间,5.7.14没有看到这样的代码。

2、第二部分用户信息

  1. # User@Host: root[root] @ localhost [] Id: 10

对应的代码:

  1. buff_len= my_snprintf(buff, 32, "%5u", thd->thread_id());
  2. if (my_b_printf(&log_file, "# User@Host: %s Id: %s\n", user_host, buff)
  3. == (uint) -1)
  4. goto err;
  5. }

user_host是一串字符串,参考代码:

  1. size_t user_host_len= (strxnmov(user_host_buff, MAX_USER_HOST_SIZE,
  2. sctx->priv_user().str
  3. ? sctx->priv_user().str : "",
  4. "[", sctx_user.length ? sctx_user.str :
  5. (thd->slave_thread ? "SQL_SLAVE" : ""),
  6. "] @ ",
  7. sctx_host.length ? sctx_host.str : "", " [",
  8. sctx_ip.length ? sctx_ip.str : "", "]",
  9. NullS) - user_host_buff);

解释如下:

  • root: m_priv_user - The user privilege we are using. May be “” for anonymous user。
  • [root]: m_user - user of the client, set to NULL until the user has been read from the connection。
  • localhost: m_host - host of the client。
  • []:client IP m_ip - client IP。
  • Id: 10 thd->thread_id()实际上就是show processlist出来的id。

    3、第三部分schema等信息

    1. # Schema: test Last_errno: 1317 Killed: 0

    对应的代码:

    1. "# Schema: %s Last_errno: %u Killed: %u\n"
    2. (thd->db().str ? thd->db().str : ""),
    3. thd->last_errno, (uint) thd->killed,
  • Schema:m_db Name of the current (default) database.If there is the current (default) database, “db” contains its name. If there is no current (default) database, “db” is NULL and “db_length” is 0. In other words, “db”, “db_length” must either be NULL, or contain a valid database name.

  • Last_errno:Variable last_errno contains the last error/warning acquired during query execution.
  • Killed: 这里代表的是终止的错误码。源码中如下:

    1. enum killed_state {
    2. NOT_KILLED=0,
    3. KILL_BAD_DATA=1,
    4. KILL_CONNECTION=ER_SERVER_SHUTDOWN,
    5. KILL_QUERY=ER_QUERY_INTERRUPTED,
    6. KILL_TIMEOUT=ER_QUERY_TIMEOUT,
    7. KILLED_NO_VALUE /* means neither of the states */
    8. };

    在错误码中代表如下:{ “ER_SERVER_SHUTDOWN”, 1053, “Server shutdown in progress” }, { “ER_QUERY_INTERRUPTED”, 1317, “Query execution was interrupted” }, { “ER_QUERY_TIMEOUT”, 1886, “Query execution was interrupted, max_statement_time exceeded” },

    4、第四部分执行信息

    这部分可能是大家最关心的部分,很多信息也是默认输出都会输出的。

    1. # Query_time: 19.254508 Lock_time: 0.001043 Rows_sent: 0 Rows_examined: 0 Rows_affected: 0
    2. # Bytes_sent: 44 Tmp_tables: 0 Tmp_disk_tables: 0 Tmp_table_sizes: 0
    3. # InnoDB_trx_id: 0

    对应代码:

    1. my_b_printf(&log_file,
    2. "# Schema: %s Last_errno: %u Killed: %u\n"
    3. "# Query_time: %s Lock_time: %s Rows_sent: %llu"
    4. " Rows_examined: %llu Rows_affected: %llu\n"
    5. "# Bytes_sent: %lu",
    6. (thd->db().str ? thd->db().str : ""),
    7. thd->last_errno, (uint) thd->killed,
    8. query_time_buff, lock_time_buff,
    9. (ulonglong) thd->get_sent_row_count(),
    10. (ulonglong) thd->get_examined_row_count(),
    11. (thd->get_row_count_func() > 0)
    12. ? (ulonglong) thd->get_row_count_func() : 0,
    13. (ulong) (thd->status_var.bytes_sent - thd->bytes_sent_old)
    14. my_b_printf(&log_file,
    15. " Tmp_tables: %lu Tmp_disk_tables: %lu "
    16. "Tmp_table_sizes: %llu",
    17. thd->tmp_tables_used, thd->tmp_tables_disk_used,
    18. thd->tmp_tables_size)
    19. snprintf(buf, 20, "%llX", thd->innodb_trx_id);及thd->innodb_trx_id
  • Query_time:语句执行的时间及实际消耗时间 。

  • Lock_time:包含MDL lock和InnoDB row lock和MyISAM表锁消耗时间的总和及锁等待消耗时间。前面已经进行了描述(实际上也并不全是锁等待的时间只是锁等待包含在其中)。

来看看Query_time和Lock_time的源码来源,它们来自于Query_logger::slow_log_write函数如下:

  1. query_utime= (current_utime > thd->start_utime) ?
  2. (current_utime - thd->start_utime) : 0;
  3. lock_utime= (thd->utime_after_lock > thd->start_utime) ?
  4. (thd->utime_after_lock - thd->start_utime) : 0;

下面是数据current_utime的来源,

  1. current_utime= thd->current_utime();

实际上就是:

  1. while (gettimeofday(&t, NULL) != 0)
  2. {}
  3. newtime= (ulonglong)t.tv_sec * 1000000 + t.tv_usec;
  4. return newtime;

获取当前时间而已
对于thd->utime_after_lock的获取已经在前文进行了描述,不再解释。

  • Rows_sent:发送给mysql客户端的行数,下面是源码中的解释 Number of rows we actually sent to the client
  • Rows_examined:InnoDB引擎层扫描的行数,下面是源码中的解释。(备注栈帧1) Number of rows read and/or evaluated for a statement. Used for slow log reporting. An examined row is defined as a row that is read and/or evaluated according to a statement condition, including increate_sort_index(). Rows may be counted more than once, e.g., a statement including ORDER BY could possibly evaluate the row in filesort() before reading it for e.g. update.
  • Rows_affected:涉及到修改的话(比如DML语句)这是受影响的行数。for DML statements: to the number of affected rows; for DDL statements: to 0.
  • Bytes_sent:发送给客户端的实际数据的字节数,它来自于 (ulong) (thd->status_var.bytes_sent - thd->bytes_sent_old)
  • Tmp_tables:临时表的个数。
  • Tmp_disk_tables:磁盘临时表的个数。
  • Tmp_table_sizes:临时表的大小。

以上三个指标来自于:

  1. thd->tmp_tables_used
  2. thd->tmp_tables_disk_used
  3. thd->tmp_tables_size

这三个指标增加的位置对应在free_tmp_table函数中如下:

  1. thd->tmp_tables_used++;
  2. if (entry->file)
  3. {
  4. thd->tmp_tables_size += entry->file->stats.data_file_length;
  5. if (entry->file->ht->db_type != DB_TYPE_HEAP)
  6. thd->tmp_tables_disk_used++;
  7. }
  • InnoDB_trx_id:事物ID,也就是trx->id,/*!< transaction id */

    5、第五部分优化器相关信息

    1. # QC_Hit: No Full_scan: No Full_join: No Tmp_table: No Tmp_table_on_disk: No
    2. # Filesort: No Filesort_on_disk: No Merge_passes: 0

    这一行来自于如下代码:

    1. my_b_printf(&log_file,
    2. "# QC_Hit: %s Full_scan: %s Full_join: %s Tmp_table: %s "
    3. "Tmp_table_on_disk: %s\n" \
    4. "# Filesort: %s Filesort_on_disk: %s Merge_passes: %lu\n",
    5. ((thd->query_plan_flags & QPLAN_QC) ? "Yes" : "No"),
    6. ((thd->query_plan_flags & QPLAN_FULL_SCAN) ? "Yes" : "No"),
    7. ((thd->query_plan_flags & QPLAN_FULL_JOIN) ? "Yes" : "No"),
    8. ((thd->query_plan_flags & QPLAN_TMP_TABLE) ? "Yes" : "No"),
    9. ((thd->query_plan_flags & QPLAN_TMP_DISK) ? "Yes" : "No"),
    10. ((thd->query_plan_flags & QPLAN_FILESORT) ? "Yes" : "No"),
    11. ((thd->query_plan_flags & QPLAN_FILESORT_DISK) ? "Yes" : "No"),

    这里注意一个处理的技巧,这里query_plan_flags中每一位都代表一个含义,这样存储既能存储足够多的信息同时存储空间也很小,是C/C++中常用的方式。

  • QC_Hit: No:是否query cache命中。

  • Full_scan:此处相当于Select_scan 的含义,是否进行了全扫描包括using index。
  • Full_join:此处相当于Select_full_join 的含义,是否被驱动表使用到了索引,如果没有使用到索引则为YES。

考虑如下的执行计划

  1. mysql> desc select *,sleep(1) from testuin a,testuin1 b where a.id1=b.id1;
  2. +----+-------------+-------+------------+------+---------------+------+---------+------+------+----------+----------------------------------------------------+
  3. | id | select_type | table | partitions | type | possible_keys | key | key_len | ref | rows | filtered | Extra |
  4. +----+-------------+-------+------------+------+---------------+------+---------+------+------+----------+----------------------------------------------------+
  5. | 1 | SIMPLE | a | NULL | ALL | NULL | NULL | NULL | NULL | 5 | 100.00 | NULL |
  6. | 1 | SIMPLE | b | NULL | ALL | NULL | NULL | NULL | NULL | 5 | 20.00 | Using where; Using join buffer (Block Nested Loop) |
  7. +----+-------------+-------+------------+------+---------------+------+---------+------+------+----------+----------------------------------------------------+
  8. 2 rows in set, 1 warning (0.00 sec)

如此输出如下:

  1. # QC_Hit: No Full_scan: Yes Full_join: Yes
  • Tmp_table:是否使用了临时表,在函数create_tmp_table中设置。
  • Tmp_table_on_disk:是否使用了磁盘临时表,如果时候innodb引擎则在create_innodb_tmp_table函数中设置。
  • Filesort:是否进行了排序,在函数filesort中设置。
  • Filesort_on_disk:是否使用了磁盘排序,同样在函数filesort中设置,但是设置之前会进行是否需要磁盘排序文件的判断。
  • Merge_passes:进行多路归并排序,归并的次数。Variable query_plan_fsort_passes collects information about file sort passes acquired during query execution.

    6、第六部分InnoDB相关信息

    1. # InnoDB_IO_r_ops: 0 InnoDB_IO_r_bytes: 0 InnoDB_IO_r_wait: 0.000000
    2. # InnoDB_rec_lock_wait: 0.000000 InnoDB_queue_wait: 0.000000
    3. # InnoDB_pages_distinct: 0

    这一行来自于如下代码:

    1. char buf[3][20];
    2. snprintf(buf[0], 20, "%.6f", thd->innodb_io_reads_wait_timer / 1000000.0);
    3. snprintf(buf[1], 20, "%.6f", thd->innodb_lock_que_wait_timer / 1000000.0);
    4. snprintf(buf[2], 20, "%.6f", thd->innodb_innodb_que_wait_timer / 1000000.0);
    5. if (my_b_printf(&log_file,
    6. "# InnoDB_IO_r_ops: %lu InnoDB_IO_r_bytes: %llu "
    7. "InnoDB_IO_r_wait: %s\n"
    8. "# InnoDB_rec_lock_wait: %s InnoDB_queue_wait: %s\n"
    9. "# InnoDB_pages_distinct: %lu\n",
    10. thd->innodb_io_reads, thd->innodb_io_read,
    11. buf[0], buf[1], buf[2], thd->innodb_page_access)
    12. == (uint) -1)
  • InnoDB_IO_r_ops:物理IO读取次数。

  • InnoDB_IO_r_bytes:物理IO读取的总字节数。
  • InnoDB_IO_r_wait:物理IO读取等待的时间。innodb 使用 BUF_IO_READ标记为物理io读取繁忙,参考函数buf_wait_for_read
  • InnoDB_rec_lock_wait:等待行锁消耗的时间。在函数que_thr_end_lock_wait中设置。
  • InnoDB_queue_wait:等待进入innodb引擎消耗的时间,在函数srv_conc_enter_innodb_with_atomics中设置。
  • InnoDB_pages_distinct:innodb访问的页数,包含物理和逻辑IO,在函数buf_page_get_gen的末尾通过_increment_page_get_statistics函数设置。

    7、第七部分set timestamp

    1. SET timestamp=1527753496;
    这一句来自源码,注意源码注释解释就是获取的服务器的当前的时间(current_utime)。 ```c / This info used to show up randomly, depending on whether the query checked the query start time or not. now we always write current timestamp to the slow log / end= my_stpcpy(end, “,timestamp=”); end= int10_to_str((long) (current_utime / 1000000), end, 10);

if (end != buff) { end++=’;’; end=’\n’; if (my_b_write(&log_file, (uchar) “SET “, 4) || my_b_write(&log_file, (uchar) buff + 1, (uint) (end-buff))) goto err; }

  1. <a name="dFnu3"></a>
  2. ## 七、总结
  3. 通过查询源码解释了一些关于MySQL慢查询的相关的知识,主要解释了慢查询是基于什么标准进行记录的,同时输出中各个指标的含义。<br />备注栈帧1:本栈帧主要跟踪Rows_examined的变化及 `join->examined_rows++;`的变化
  4. ```c
  5. (gdb) info b
  6. Num Type Disp Enb Address What
  7. 1 breakpoint keep y 0x0000000000ebd5f3 in main(int, char**) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/main.cc:25
  8. breakpoint already hit 1 time
  9. 4 breakpoint keep y 0x000000000155b94f in do_select(JOIN*) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:872
  10. breakpoint already hit 5 times
  11. 5 breakpoint keep y 0x000000000155ca39 in evaluate_join_record(JOIN*, QEP_TAB*) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:1473
  12. breakpoint already hit 20 times
  13. 6 breakpoint keep y 0x00000000019b4313 in ha_innobase::index_first(uchar*)
  14. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9547
  15. breakpoint already hit 4 times
  16. 7 breakpoint keep y 0x00000000019b45cd in ha_innobase::rnd_next(uchar*)
  17. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9651
  18. 8 breakpoint keep y 0x00000000019b2ba6 in ha_innobase::index_read(uchar*, uchar const*, uint, ha_rkey_function)
  19. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9004
  20. breakpoint already hit 3 times
  21. 9 breakpoint keep y 0x00000000019b4233 in ha_innobase::index_next(uchar*)
  22. at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9501
  23. breakpoint already hit 5 times
  24. #0 ha_innobase::index_next (this=0x7fff2cbc6b40, buf=0x7fff2cbc7080 "\375\n") at /root/mysql5.7.14/percona-server-5.7.14-7/storage/innobase/handler/ha_innodb.cc:9501
  25. #1 0x0000000000f680d8 in handler::ha_index_next (this=0x7fff2cbc6b40, buf=0x7fff2cbc7080 "\375\n") at /root/mysql5.7.14/percona-server-5.7.14-7/sql/handler.cc:3269
  26. #2 0x000000000155fa02 in join_read_next (info=0x7fff2c007750) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:2660
  27. #3 0x000000000155c397 in sub_select (join=0x7fff2c007020, qep_tab=0x7fff2c007700, end_of_records=false)
  28. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:1274
  29. #4 0x000000000155bd06 in do_select (join=0x7fff2c007020) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:944
  30. #5 0x0000000001559bdc in JOIN::exec (this=0x7fff2c007020) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_executor.cc:199
  31. #6 0x00000000015f9ea6 in handle_query (thd=0x7fff2c000b70, lex=0x7fff2c003150, result=0x7fff2c006cd0, added_options=0, removed_options=0)
  32. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_select.cc:184
  33. #7 0x00000000015acd05 in execute_sqlcom_select (thd=0x7fff2c000b70, all_tables=0x7fff2c006688) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:5391
  34. #8 0x00000000015a5320 in mysql_execute_command (thd=0x7fff2c000b70, first_level=true) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:2889
  35. #9 0x00000000015adcd6 in mysql_parse (thd=0x7fff2c000b70, parser_state=0x7ffff035b600) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:5836
  36. #10 0x00000000015a1b95 in dispatch_command (thd=0x7fff2c000b70, com_data=0x7ffff035bd70, command=COM_QUERY)
  37. at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:1447
  38. #11 0x00000000015a09c6 in do_command (thd=0x7fff2c000b70) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/sql_parse.cc:1010
  39. #12 0x00000000016e29d0 in handle_connection (arg=0x3859ae0) at /root/mysql5.7.14/percona-server-5.7.14-7/sql/conn_handler/connection_handler_per_thread.cc:312
  40. #13 0x0000000001d7bfdc in pfs_spawn_thread (arg=0x38607b0) at /root/mysql5.7.14/percona-server-5.7.14-7/storage/perfschema/pfs.cc:2188
  41. #14 0x0000003f74807aa1 in start_thread () from /lib64/libpthread.so.0
  42. #15 0x0000003f740e8bcd in clone () from /lib64/libc.so.6