Hi, Nikita,
this is a review of the combined diff of your branch.
Of `git diff github/10.8...github/bb-10.8-online-alter`, that is of
cf7cc376bac fix savepoints in myisam
4a128ca5458 fix main.delayed
7300151cc84 fix skipping rocksdb
19504dfdbcc add binlog/standalone combinations
55e993595aa add rocksdb combination
06645b17450 savepoints
9ffcd13b9fe rename tests
730926ae962 MDEV-16329 [5/5] ALTER ONLINE TABLE
cf52ea33a70 MDEV-16329 [4/5] Refactor MYSQL_BIN_LOG: extract Event_log ancestor
0445c8f1ee9 MDEV-16329 [3/5] use binlog_cache_data directly in most places
3f3da0f65a3 MDEV-16329 [2/5] refactor binlog and cache_mngr
dc79667ca73 MDEV-16329 [1/5] add THD::binlog_get_cache_mngr
6f9dd00ce2d rpl: repack table_def
faf8680df8d Copy_field: add const to arguments
It was mostly fine the last time already, so only few comments below:
> diff --git a/mysql-test/include/have_log_bin_off.require b/mysql-test/include/have_log_bin_off.require
> new file mode 100644
> index 00000000000..979dbe75f80
> --- /dev/null
> +++ b/mysql-test/include/have_log_bin_off.require
> @@ -0,0 +1,2 @@
> +Variable_name Value
> +log_bin OFF
better use `if (...) { skip }` pattern
*.require is an obsolete 20-year old technique from before if- and skip-times
> diff --git a/mysql-test/main/alter_table.test b/mysql-test/main/alter_table.test
> index 31c69783248..91423ee8c1f 100644
> --- a/mysql-test/main/alter_table.test
> +++ b/mysql-test/main/alter_table.test
> @@ -1532,8 +1532,12 @@ ALTER TABLE t1 DROP INDEX i1, DROP INDEX i2, DROP INDEX i3, DROP INDEX i4;
> ALTER TABLE t1 ADD INDEX i1(b), ALGORITHM= INPLACE, LOCK= NONE;
> ALTER TABLE t1 ADD INDEX i2(b), ALGORITHM= INPLACE, LOCK= SHARED;
> ALTER TABLE t1 ADD INDEX i3(b), ALGORITHM= INPLACE, LOCK= EXCLUSIVE;
> ---error ER_ALTER_OPERATION_NOT_SUPPORTED_REASON
> +--disable_info
> +--disable_warnings
> +--error 0,ER_ALTER_OPERATION_NOT_SUPPORTED_REASON
why, because of embedded?
> ALTER TABLE t1 ADD INDEX i4(b), ALGORITHM= COPY, LOCK= NONE;
> +--enable_warnings
> +--enable_info
> ALTER TABLE t1 ADD INDEX i5(b), ALGORITHM= COPY, LOCK= SHARED;
> ALTER TABLE t1 ADD INDEX i6(b), ALGORITHM= COPY, LOCK= EXCLUSIVE;
>
> diff --git a/mysql-test/main/alter_table_locknone.result b/mysql-test/main/alter_table_locknone.result
> new file mode 100644
> index 00000000000..ea040925efe
> --- /dev/null
> +++ b/mysql-test/main/alter_table_locknone.result
> @@ -0,0 +1,270 @@
> +create table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b')) engine=myisam;
> +insert into t1 (a) values (1),(2),(3);
> +alter online table t1 modify b int default 5, alter c set default 'X';
> +alter online table t1 change b new_name int;
> +alter online table t1 modify e enum('a','b','c');
> +alter online table t1 comment "new comment";
> +alter table t1 add constraint q check (a > 0);
> +alter online table t1 drop constraint q;
> +alter online table t1 algorithm=INPLACE, lock=NONE;
> +alter online table t1;
> +alter table t1 algorithm=INPLACE;
> +alter table t1 lock=NONE;
> +show create table t1;
> +Table Create Table
> +t1 CREATE TABLE `t1` (
> + `a` int(11) NOT NULL,
> + `new_name` int(11) DEFAULT NULL,
> + `c` varchar(80) DEFAULT 'X',
> + `e` enum('a','b','c') DEFAULT NULL,
> + PRIMARY KEY (`a`)
> +) ENGINE=MyISAM DEFAULT CHARSET=latin1 COMMENT='new comment'
> +drop table t1;
> +create temporary table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b'));
> +insert into t1 (a) values (1),(2),(3);
> +alter online table t1 modify b int default 5, alter c set default 'X';
> +alter online table t1 change b new_name int;
> +alter online table t1 modify e enum('a','b','c');
> +alter online table t1 comment "new comment";
> +alter online table t1 rename to t2;
> +show create table t2;
> +Table Create Table
> +t2 CREATE TEMPORARY TABLE `t2` (
> + `a` int(11) NOT NULL,
> + `new_name` int(11) DEFAULT NULL,
> + `c` varchar(80) DEFAULT 'X',
> + `e` enum('a','b','c') DEFAULT NULL,
> + PRIMARY KEY (`a`)
> +) ENGINE=MyISAM DEFAULT CHARSET=latin1 COMMENT='new comment'
> +drop table t2;
> +create table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b')) engine=aria;
> +insert into t1 (a) values (1),(2),(3);
> +alter online table t1 modify b int default 5;
> +alter online table t1 change b new_name int;
> +alter online table t1 modify e enum('a','b','c');
> +alter online table t1 comment "new comment";
> +show create table t1;
> +Table Create Table
> +t1 CREATE TABLE `t1` (
> + `a` int(11) NOT NULL,
> + `new_name` int(11) DEFAULT NULL,
> + `c` varchar(80) DEFAULT NULL,
> + `e` enum('a','b','c') DEFAULT NULL,
> + PRIMARY KEY (`a`)
> +) ENGINE=Aria DEFAULT CHARSET=latin1 PAGE_CHECKSUM=1 COMMENT='new comment'
> +alter online table t1 page_checksum=1;
> +alter online table t1 page_checksum=0;
> +ERROR 0A000: LOCK=NONE is not supported for this operation. Try LOCK=SHARED
Why LOCK=NONE would be not supported now?
Shouldn't your new feature allow just everything with LOCK=NONE?
> +drop table t1;
> +create table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b'));
> +insert into t1 (a) values (1),(2),(3);
> +alter online table t1 drop column b, add b int;
> +ERROR 0A000: LOCK=NONE is not supported for this operation. Try LOCK=SHARED
...
> diff --git a/mysql-test/main/alter_table_online.combinations b/mysql-test/main/alter_table_online.combinations
> new file mode 100644
> index 00000000000..ae144432c68
> --- /dev/null
> +++ b/mysql-test/main/alter_table_online.combinations
> @@ -0,0 +1,2 @@
> +[innodb]
> +[rocksdb]
what about non-trans tables?
> diff --git a/mysql-test/main/alter_table_online.test b/mysql-test/main/alter_table_online.test
> index 6ef9661c43b..815483d252d 100644
> --- a/mysql-test/main/alter_table_online.test
> +++ b/mysql-test/main/alter_table_online.test
> @@ -1,335 +1,596 @@
> -#
> -# Test of ALTER ONLINE TABLE syntax
> -#
> -
> ---source include/have_innodb.inc
> ---source include/have_partition.inc
> -#
> -# Test of things that can be done online
> -#
> -
> -create table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b')) engine=myisam;
> -insert into t1 (a) values (1),(2),(3);
> -
> -alter online table t1 modify b int default 5, alter c set default 'X';
> -alter online table t1 change b new_name int;
> -alter online table t1 modify e enum('a','b','c');
> -alter online table t1 comment "new comment";
> -alter table t1 add constraint q check (a > 0);
> -alter online table t1 drop constraint q;
> -
> -# No OPs
> -
> -alter online table t1 algorithm=INPLACE, lock=NONE;
> -alter online table t1;
> -alter table t1 algorithm=INPLACE;
> -alter table t1 lock=NONE;
> -show create table t1;
> -drop table t1;
> +-- source include/have_debug_sync.inc
> +-- source include/not_embedded.inc
> +-- source alter_table_online_binlog.inc
>
> -#
> -# everything with temporary tables is "online", i.e. without locks
> -#
> -create temporary table t1 (a int not null primary key, b int, c varchar(80), e enum('a','b'));
> -insert into t1 (a) values (1),(2),(3);
> -
> -alter online table t1 modify b int default 5, alter c set default 'X';
> -alter online table t1 change b new_name int;
> -alter online table t1 modify e enum('a','b','c');
> -alter online table t1 comment "new comment";
> -alter online table t1 rename to t2;
> -show create table t2;
> -drop table t2;
> +-- disable_query_log
> +-- if ($MTR_COMBINATION_INNODB) {
> +-- source include/have_innodb.inc
> +set default_storage_engine= innodb;
> +-- }
> +-- if ($MTR_COMBINATION_ROCKSDB) {
> +-- source include/have_rocksdb.inc
> +set default_storage_engine= rocksdb;
> +-- }
> +-- enable_query_log
> +
> +-- let $default_engine= `select @@default_storage_engine`
> +
> +--connect (con2, localhost, root,,)
> +--connection default
> +
> +
> +--echo #
> +--echo # Test insert
> +--echo #
> +
> +--echo # Insert and add column
> +create or replace table t1 (a int) engine=innodb;
why do you specify the engine explicitly if you run it in two engine
combinations?
> +insert t1 values (5);
> +
> +--connection con2
> +--send
> +set debug_sync= 'now WAIT_FOR ended';
> +
> +--connection default
> +set debug_sync= 'alter_table_copy_end SIGNAL ended WAIT_FOR end';
> +
...
> +
> +--connection default
> +--reap
> +--sorted_result
> +select * from t1;
> +
> +--echo #
> +--echo # MYISAM. Only Inserts can be tested.
why?
> +--echo #
> +
> +create or replace table t1 (a int) engine=myisam;
> +insert t1 values (5);
> +
> +--connection con2
> +--send
> +set debug_sync= 'now WAIT_FOR ended';
> +
> +--connection default
> +set debug_sync= 'alter_table_copy_end SIGNAL ended WAIT_FOR end';
> +
> +--send
> +alter table t1 add b int NULL, algorithm= copy, lock= none;
> +
...
> diff --git a/mysql-test/main/mdl_sync.test b/mysql-test/main/mdl_sync.test
> index 3df19aca806..8c6bc76f5ca 100644
> --- a/mysql-test/main/mdl_sync.test
> +++ b/mysql-test/main/mdl_sync.test
> @@ -42,7 +42,7 @@ connection con1;
> set debug_sync='mdl_upgrade_lock SIGNAL parked WAIT_FOR go';
> --send alter table t1 rename t3
>
> -connection default;
> + connection default;
eh? (here and many other changes in this file)
> --echo connection: default
> set debug_sync= 'now WAIT_FOR parked';
>
> diff --git a/mysql-test/suite/gcol/inc/gcol_select.inc b/mysql-test/suite/gcol/inc/gcol_select.inc
> index 2386c55fdbc..8f6314233bd 100644
> --- a/mysql-test/suite/gcol/inc/gcol_select.inc
> +++ b/mysql-test/suite/gcol/inc/gcol_select.inc
> @@ -872,7 +872,9 @@ CREATE TABLE t1(a INT);
> INSERT INTO t1 VALUES(2147483647);
> ALTER TABLE t1 ADD COLUMN h INT AS (a) VIRTUAL;
> ALTER TABLE t1 CHANGE h i INT AS (a) VIRTUAL, ALGORITHM=COPY;
> +--error ER_WARN_DATA_OUT_OF_RANGE,ER_ALTER_OPERATION_NOT_SUPPORTED_REASON
why two? embedded, again?
> ALTER TABLE t1 ADD COLUMN b SMALLINT AS (a) VIRTUAL, ALGORITHM=COPY, LOCK=NONE;
> +--error ER_WARN_DATA_OUT_OF_RANGE,ER_ALTER_OPERATION_NOT_SUPPORTED_REASON
> ALTER TABLE t1 ADD COLUMN e SMALLINT AS (a) VIRTUAL, ALGORITHM=COPY, LOCK=NONE;
> ALTER TABLE t1 ADD COLUMN f SMALLINT AS (a) VIRTUAL, ALGORITHM=COPY, LOCK=SHARED;
> ALTER TABLE t1 ADD COLUMN g SMALLINT AS (a) VIRTUAL, ALGORITHM=COPY, LOCK=EXCLUSIVE;
> diff --git a/sql/handler.h b/sql/handler.h
> index fe61666bf20..5e46fa59d7e 100644
> --- a/sql/handler.h
> +++ b/sql/handler.h
> @@ -3538,6 +3543,7 @@ class handler :public Sql_alloc
> /** to be actually called to get 'check()' functionality*/
> int ha_check(THD *thd, HA_CHECK_OPT *check_opt);
> int ha_repair(THD* thd, HA_CHECK_OPT* check_opt);
> + virtual void open_read_view(){}
why did you change to that from start_consistent_snapshot()?
> void ha_start_bulk_insert(ha_rows rows, uint flags= 0)
> {
> DBUG_ENTER("handler::ha_start_bulk_insert");
> diff --git a/sql/log.cc b/sql/log.cc
> index 70ceecc66f8..9beab80773c 100644
> --- a/sql/log.cc
> +++ b/sql/log.cc
> @@ -101,6 +102,9 @@ static int binlog_flush_cache(THD *thd, binlog_cache_mngr *cache_mngr,
> Log_event *end_ev, bool all, bool using_stmt,
> bool using_trx, bool is_ro_1pc);
>
> +int binlog_online_alter_commit(THD *thd, bool all);
> +void binlog_online_alter_rollback(THD *thd, bool all);
static?
> +
> static const LEX_CSTRING write_error_msg=
> { STRING_WITH_LEN("error writing to the binary log") };
>
> @@ -3565,6 +3613,91 @@ bool MYSQL_BIN_LOG::open_index_file(const char *index_file_name_arg,
> }
>
>
> +bool Event_log::open(const char *log_name,
> + const char *new_name, ulong next_file_number,
> + enum cache_type io_cache_type_arg)
> +{
> + bool error= false;
> + if (log_name || new_name)
> + {
> + error= MYSQL_LOG::open(
> +#ifdef HAVE_PSI_INTERFACE
> + 0,
> +#endif
> + log_name, LOG_NORMAL, new_name, next_file_number, io_cache_type_arg);
> + }
> + else
> + {
> +#ifdef HAVE_PSI_INTERFACE
> + /* Keep the key for reopen */
> + m_log_file_key= 0;
> +#endif
> + error= init_io_cache(&log_file, -1, LOG_BIN_IO_SIZE,
> + io_cache_type_arg, 0, 0,
> + MYF(MY_WME | MY_NABP | MY_WAIT_IF_FULL));
> +
> + log_state= LOG_OPENED;
> + inited= true;
> + }
> + if (error)
> + return error;
> +
> + longlong bytes_written= write_description_event(
> + (enum_binlog_checksum_alg)binlog_checksum_options,
> + encrypt_binlog, false);
> + error= bytes_written < 0;
> + return error;
> +}
> +
> +longlong
> +Event_log::write_description_event(enum_binlog_checksum_alg checksum_alg,
> + bool encrypt, bool dont_set_created)
> +{
> + Format_description_log_event s(BINLOG_VERSION);
> + /*
> + don't set LOG_EVENT_BINLOG_IN_USE_F for SEQ_READ_APPEND io_cache
> + as we won't be able to reset it later
> + */
> + if (io_cache_type == WRITE_CACHE)
this was incorrectly rebased. in the previous patch you *moved*
this block of code from MYSQL_BIN_LOG::open into a separate method.
here you copied it. you forgot to delete it from MYSQL_BIN_LOG::open().
> + s.flags |= LOG_EVENT_BINLOG_IN_USE_F;
> + s.checksum_alg= checksum_alg;
> +
> + crypto.scheme = 0;
> + DBUG_ASSERT(s.checksum_alg != BINLOG_CHECKSUM_ALG_UNDEF);
> + if (!s.is_valid())
> + return -1;
> + s.dont_set_created= dont_set_created;
> + if (write_event(&s, 0, &log_file))
> + return -1;
> +
> + if (encrypt)
> + {
> + uint key_version= encryption_key_get_latest_version(ENCRYPTION_KEY_SYSTEM_DATA);
> + if (key_version == ENCRYPTION_KEY_VERSION_INVALID)
> + {
> + sql_print_error("Failed to enable encryption of binary logs");
> + return -1;
> + }
> +
> + if (key_version != ENCRYPTION_KEY_NOT_ENCRYPTED)
> + {
> + if (my_random_bytes(crypto.nonce, sizeof(crypto.nonce)))
> + return -1;
> +
> + Start_encryption_log_event sele(1, key_version, crypto.nonce);
> + sele.checksum_alg= s.checksum_alg;
> + if (write_event(&sele, 0, &log_file))
> + return -1;
> +
> + // Start_encryption_log_event is written, enable the encryption
> + if (crypto.init(sele.crypto_scheme, key_version))
> + return -1;
> + }
> + }
> + return (longlong)s.data_written;
> +}
> +
> +
> /**
> Open a (new) binlog file.
>
> @@ -7252,6 +7477,160 @@ class CacheWriter: public Log_event_writer
> bool first;
> };
>
> +int cache_copy(IO_CACHE *to, IO_CACHE *from)
> +{
> + DBUG_ENTER("cache_copy");
> + if (reinit_io_cache(from, READ_CACHE, 0, 0, 0))
> + DBUG_RETURN(ER_ERROR_ON_WRITE);
> + size_t bytes_in_cache= my_b_bytes_in_cache(from);
> +
> + do
> + {
> + my_b_write(to, from->read_pos, bytes_in_cache);
> +
> + from->read_pos += bytes_in_cache;
> + bytes_in_cache= my_b_fill(from);
> + if (from->error || to->error)
> + DBUG_RETURN(ER_ERROR_ON_WRITE);
> + } while (bytes_in_cache);
> +
> + DBUG_RETURN(0);
> +}
> +
> +int binlog_online_alter_commit(THD *thd, bool all)
> +{
> + DBUG_ENTER("online_alter_commit");
> + int error= 0;
> +#ifdef HAVE_REPLICATION
> +
> + if (thd->online_alter_cache_list.empty())
> + DBUG_RETURN(0);
> +
> + bool is_ending_transaction= ending_trans(thd, all);
> +
> + for (auto &cache_mngr: thd->online_alter_cache_list)
> + {
> + auto *binlog= cache_mngr.share->online_alter_binlog;
> + DBUG_ASSERT(binlog);
> +
> + error= binlog_flush_pending_rows_event(thd,
> + /*
> + do not set STMT_END for last event
> + to leave table open in altering thd
> + */
> + false,
> + true,
> + binlog,
> + is_ending_transaction
> + ? &cache_mngr.trx_cache
> + : &cache_mngr.stmt_cache);
> + if (is_ending_transaction)
> + {
> + mysql_mutex_lock(binlog->get_log_lock());
> + error= binlog->write_cache(thd, &cache_mngr.trx_cache.cache_log);
> +
> + mysql_mutex_unlock(binlog->get_log_lock());
> + }
> + else
> + {
> + error= cache_copy(&cache_mngr.trx_cache.cache_log,
> + &cache_mngr.stmt_cache.cache_log);
Why do you need two caches here? It seems you could have just one trx_cache
and if the statement fails you simply truncate it to the beginning of
the statement.
> + }
> +
> + if (error)
> + {
> + my_error(ER_ERROR_ON_WRITE, MYF(ME_ERROR_LOG),
> + binlog->get_name(), errno);
> + binlog_online_alter_cleanup(thd->online_alter_cache_list,
> + is_ending_transaction);
> + DBUG_RETURN(error);
> + }
> + }
> +
> + binlog_online_alter_cleanup(thd->online_alter_cache_list,
> + is_ending_transaction);
> +
> + for (TABLE *table= thd->open_tables; table; table= table->next)
> + {
> + table->online_alter_cache= NULL;
Why?
> + }
> +#endif // HAVE_REPLICATION
> + DBUG_RETURN(error);
> +}
> +
> +void binlog_online_alter_rollback(THD *thd, bool all)
> +{
> +#ifdef HAVE_REPLICATION
> + bool is_ending_trans= ending_trans(thd, all);
> +
> + /*
> + This is a crucial moment that we are running through
> + thd->online_alter_cache_list, and not through thd->open_tables to cleanup
> + stmt cache, though both have it. The reason is that the tables can be closed
> + to that moment in case of an error.
> + The same reason applies to the fact we don't store cache_mngr in the table
> + itself -- because it can happen to be not existing.
> + Still in case if tables are left opened
> + */
> + binlog_online_alter_cleanup(thd->online_alter_cache_list, is_ending_trans);
> +#endif // HAVE_REPLICATION
> +}
> +
> +SAVEPOINT** find_savepoint_in_list(THD *thd, LEX_CSTRING name,
> + SAVEPOINT ** const list);
> +
> +SAVEPOINT* savepoint_add(THD *thd, LEX_CSTRING name, SAVEPOINT **list,
> + int (*release_old)(THD*, SAVEPOINT*));
may be in transaction.h?
> +
> +int online_alter_savepoint_set(THD *thd, LEX_CSTRING name)
> +{
> +
> + DBUG_ENTER("binlog_online_alter_savepoint");
> +#ifdef HAVE_REPLICATION
> + if (thd->online_alter_cache_list.empty())
> + DBUG_RETURN(0);
> +
> + if (savepoint_alloc_size < sizeof (SAVEPOINT) + sizeof(my_off_t))
> + savepoint_alloc_size= sizeof (SAVEPOINT) + sizeof(my_off_t);
> +
> + for (auto &cache: thd->online_alter_cache_list)
> + {
> + if (cache.share->db_type()->savepoint_set == NULL)
> + continue;
> +
> + SAVEPOINT *sv= savepoint_add(thd, name, &cache.sv_list, NULL);
wouldn't it be simpler to log SAVEPOINT and ROLLBACK/RELEASE
as Query_log_event to the cache? Then you won't need to do
any maintenance.
> + if(unlikely(sv == NULL))
> + DBUG_RETURN(1);
> + my_off_t *pos= (my_off_t*)(sv+1);
> + *pos= cache.trx_cache.get_byte_position();
> +
> + sv->prev= cache.sv_list;
> + cache.sv_list= sv;
> + }
> +#endif
> + DBUG_RETURN(0);
> +}
> +
> +int online_alter_savepoint_rollback(THD *thd, LEX_CSTRING name)
> +{
> + DBUG_ENTER("online_alter_savepoint_rollback");
> +#ifdef HAVE_REPLICATION
> + for (auto &cache: thd->online_alter_cache_list)
> + {
> + if (cache.share->db_type()->savepoint_set == NULL)
> + continue;
> +
> + SAVEPOINT **sv= find_savepoint_in_list(thd, name, &cache.sv_list);
> + // sv is null if savepoint was set up before online table was modified
> + my_off_t pos= *sv ? *(my_off_t*)(*sv+1) : 0;
> +
> + cache.trx_cache.restore_savepoint(pos);
> + }
> +
> +#endif
> + DBUG_RETURN(0);
> +}
> +
> /*
> Write the contents of a cache to the binary log.
>
> diff --git a/sql/rpl_rli.h b/sql/rpl_rli.h
> index cc807852bf2..57757ecf3dd 100644
> --- a/sql/rpl_rli.h
> +++ b/sql/rpl_rli.h
> @@ -900,14 +900,20 @@ struct rpl_group_info
> }
> }
>
> - bool get_table_data(TABLE *table_arg, table_def **tabledef_var, TABLE **conv_table_var) const
> + bool get_table_data(TABLE *table_arg, table_def **tabledef_var,
> + TABLE **conv_table_var,
> + const Copy_field *copy[], const Copy_field **copy_end) const
> {
> DBUG_ASSERT(tabledef_var && conv_table_var);
> for (TABLE_LIST *ptr= tables_to_lock ; ptr != NULL ; ptr= ptr->next_global)
> if (ptr->table == table_arg)
> {
> - *tabledef_var= &static_cast<RPL_TABLE_LIST*>(ptr)->m_tabledef;
> - *conv_table_var= static_cast<RPL_TABLE_LIST*>(ptr)->m_conv_table;
> + auto *rpl_table_list= static_cast<RPL_TABLE_LIST*>(ptr);
> + if (rpl_table_list->m_tabledef_valid)
when can it be false?
> + *tabledef_var= &rpl_table_list->m_tabledef;
> + *conv_table_var= rpl_table_list->m_conv_table;
> + *copy= rpl_table_list->m_online_alter_copy_fields;
> + *copy_end= rpl_table_list->m_online_alter_copy_fields_end;
> DBUG_PRINT("debug", ("Fetching table data for table %s.%s:"
> " tabledef: %p, conv_table: %p",
> table_arg->s->db.str, table_arg->s->table_name.str,
> diff --git a/sql/sql_table.cc b/sql/sql_table.cc
> index b07efb29bba..9dfe2178667 100644
> --- a/sql/sql_table.cc
> +++ b/sql/sql_table.cc
> @@ -9533,6 +9539,19 @@ bool mysql_alter_table(THD *thd, const LEX_CSTRING *new_db,
> has been already processed.
> */
> table_list->required_type= TABLE_TYPE_NORMAL;
> +
> +
> + if (alter_info->requested_lock == Alter_info::ALTER_TABLE_LOCK_SHARED
> + || alter_info->requested_lock == Alter_info::ALTER_TABLE_LOCK_EXCLUSIVE
> + || thd->locked_tables_mode == LTM_LOCK_TABLES
> + || thd->lex->sql_command == SQLCOM_OPTIMIZE
> + || alter_info->algorithm(thd) == Alter_info::ALTER_TABLE_ALGORITHM_NOCOPY)
why only NOCOPY? What about INPLACE and INSTANT?
the rest of the condition looks fine.
> + online= false;
> +
> + if (online)
> + {
> + table_list->lock_type= TL_READ;
> + }
>
> DEBUG_SYNC(thd, "alter_table_before_open_tables");
>
> @@ -10961,6 +10983,58 @@ bool mysql_trans_commit_alter_copy_data(THD *thd)
> DBUG_RETURN(error);
> }
>
> +#ifdef HAVE_REPLICATION
> +static int online_alter_read_from_binlog(THD *thd, rpl_group_info *rgi,
> + Cache_flip_event_log *log)
> +{
> + MEM_ROOT event_mem_root;
> + Query_arena backup_arena;
> + Query_arena event_arena(&event_mem_root, Query_arena::STMT_INITIALIZED);
> + init_sql_alloc(key_memory_gdl, &event_mem_root,
> + MEM_ROOT_BLOCK_SIZE, 0, MYF(0));
> +
> + int error= 0;
> +
> + IO_CACHE *log_file= log->flip();
> +
> + thd_progress_report(thd, 0, my_b_write_tell(log_file));
> +
> + Abort_on_warning_instant_set old_abort_on_warning(thd, 0);
> + do
> + {
> + const auto *descr_event= rgi->rli->relay_log.description_event_for_exec;
> + auto *ev= Log_event::read_log_event(log_file, descr_event, false);
> + if (!ev)
> + break;
> +
> + ev->thd= thd;
> + thd->set_n_backup_active_arena(&event_arena, &backup_arena);
> + error= ev->apply_event(rgi);
> + thd->restore_active_arena(&event_arena, &backup_arena);
> +
> + event_arena.free_items();
> + free_root(&event_mem_root, MYF(MY_KEEP_PREALLOC));
> + if (ev != rgi->rli->relay_log.description_event_for_exec)
> + delete ev;
> + thd_progress_report(thd, my_b_tell(log_file), thd->progress.max_counter);
> + DEBUG_SYNC(thd, "alter_table_online_progress");
> + } while(!error);
> +
> + return error;
> +}
> +#endif // HAVE_REPLICATION
> +
> +static void online_alter_cleanup_binlog(THD *thd, TABLE_SHARE *s)
> +{
> +#ifdef HAVE_REPLICATION
> + if (!s->online_alter_binlog)
> + return;
> + // s->online_alter_binlog->reset_logs(thd, false, NULL, 0, 0);
forgot to delete?
> + s->online_alter_binlog->cleanup();
> + s->online_alter_binlog->~Cache_flip_event_log();
> + s->online_alter_binlog= NULL;
> +#endif
> +}
>
> static int
> copy_data_between_tables(THD *thd, TABLE *from, TABLE *to,
> @@ -11306,6 +11431,76 @@ copy_data_between_tables(THD *thd, TABLE *from, TABLE *to,
> cleanup_done= 1;
> to->file->extra(HA_EXTRA_NO_IGNORE_DUP_KEY);
>
> +#ifdef HAVE_REPLICATION
> + if (likely(online && error < 0))
> + {
> + Ha_trx_info *trx_info_save= thd->transaction->all.ha_list;
> + thd->transaction->all.ha_list = NULL;
why?
> + thd_progress_next_stage(thd);
> + Table_map_log_event table_event(thd, from, from->s->table_map_id,
> + from->file->has_transactions());
> + Relay_log_info rli(false);
> + rpl_group_info rgi(&rli);
> + RPL_TABLE_LIST rpl_table(to, TL_WRITE, from, table_event.get_table_def(),
> + copy, copy_end);
> + Cache_flip_event_log *binlog= from->s->online_alter_binlog;
> + rgi.thd= thd;
> + rgi.tables_to_lock= &rpl_table;
> +
> + rgi.m_table_map.set_table(from->s->table_map_id, to);
> +
> + DBUG_ASSERT(binlog->is_open());
> +
> + rli.relay_log.description_event_for_exec=
> + new Format_description_log_event(4);
> +
> + // We restore bitmaps, because update event is going to mess up with them.
> + to->default_column_bitmaps();
> +
> + error= online_alter_read_from_binlog(thd, &rgi, binlog);
> +
> + DEBUG_SYNC(thd, "alter_table_online_before_lock");
> +
> + int lock_error=
> + thd->mdl_context.upgrade_shared_lock(from->mdl_ticket, MDL_EXCLUSIVE,
> + (double)thd->variables.lock_wait_timeout);
> + if (!error)
> + error= lock_error;
> +
> + if (!error)
> + {
> + thd_progress_next_stage(thd);
> + error= online_alter_read_from_binlog(thd, &rgi, binlog);
> + }
> +
> + thd->transaction->all.ha_list = trx_info_save;
> + }
> + else if (unlikely(online)) // error was on copy stage
> + {
> + /*
> + We need to issue a barrier to clean up gracefully.
> + Without this, following possible:
> + T1: ALTER TABLE starts
> + T2: INSERT starts
> + T1: ALTER TABLE fails with error (i.e. ER_DUP_KEY)
> + T1: from->s->online_alter_binlog sets to NULL
> + T2: INSERT committs
> + T2: thd->online_alter_cache_list is not empty
> + T2: binlog_commit: DBUG_ASSERT(binlog); is issued.
1. do you have a test for that?
2. can it be fixed, like, on iterating thd->online_alter_cache_list
thd notices that from->s->online_alter_cache_list is NULL, so
it simply discards the cache? Then you won't need MDL_SHARED_NO_WRITE
> + */
> + // Ignore the return result. We already have an error.
> + thd->mdl_context.upgrade_shared_lock(from->mdl_ticket,
> + MDL_SHARED_NO_WRITE,
> + thd->variables.lock_wait_timeout);
> + }
> +#endif
> +
> + if (error > 0 && !from->s->tmp_table)
> + {
> + /* We are going to drop the temporary table */
> + to->file->extra(HA_EXTRA_PREPARE_FOR_DROP);
> + }
> +
> DEBUG_SYNC(thd, "copy_data_between_tables_before_reset_backup_lock");
> if (backup_reset_alter_copy_lock(thd))
> error= 1;
Regards,
Sergei
VP of MariaDB Server Engineering
and security(a)mariadb.org