1661 lines
76 KiB
Go
1661 lines
76 KiB
Go
// Copyright 2019 PingCAP, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package ddl_test
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"math"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
. "github.com/pingcap/check"
|
|
"github.com/pingcap/errors"
|
|
"github.com/pingcap/failpoint"
|
|
"github.com/pingcap/parser/model"
|
|
"github.com/pingcap/parser/mysql"
|
|
"github.com/pingcap/tidb/config"
|
|
"github.com/pingcap/tidb/ddl"
|
|
ddlutil "github.com/pingcap/tidb/ddl/util"
|
|
"github.com/pingcap/tidb/domain"
|
|
"github.com/pingcap/tidb/errno"
|
|
"github.com/pingcap/tidb/infoschema"
|
|
"github.com/pingcap/tidb/kv"
|
|
"github.com/pingcap/tidb/meta"
|
|
"github.com/pingcap/tidb/meta/autoid"
|
|
"github.com/pingcap/tidb/planner/core"
|
|
"github.com/pingcap/tidb/session"
|
|
"github.com/pingcap/tidb/sessionctx"
|
|
"github.com/pingcap/tidb/sessionctx/variable"
|
|
"github.com/pingcap/tidb/store/mockstore"
|
|
"github.com/pingcap/tidb/tablecodec"
|
|
"github.com/pingcap/tidb/util/admin"
|
|
"github.com/pingcap/tidb/util/collate"
|
|
"github.com/pingcap/tidb/util/gcutil"
|
|
"github.com/pingcap/tidb/util/mock"
|
|
"github.com/pingcap/tidb/util/testkit"
|
|
. "github.com/pingcap/tidb/util/testutil"
|
|
"github.com/tikv/client-go/v2/mockstore/cluster"
|
|
)
|
|
|
|
// Make it serial because config is modified in test cases.
|
|
var _ = SerialSuites(&testSerialSuite{})
|
|
|
|
// TODO(tangenta): Move all the parallel tests out of this file.
|
|
var _ = Suite(&testIntegrationSuite7{&testIntegrationSuite{}})
|
|
|
|
type testSerialSuite struct {
|
|
CommonHandleSuite
|
|
store kv.Storage
|
|
cluster cluster.Cluster
|
|
dom *domain.Domain
|
|
}
|
|
|
|
func (s *testSerialSuite) SetUpSuite(c *C) {
|
|
session.SetSchemaLease(200 * time.Millisecond)
|
|
session.DisableStats4Test()
|
|
config.UpdateGlobal(func(conf *config.Config) {
|
|
// Update config here.
|
|
})
|
|
|
|
ddl.SetWaitTimeWhenErrorOccurred(1 * time.Microsecond)
|
|
|
|
var err error
|
|
s.store, err = mockstore.NewMockStore(
|
|
mockstore.WithClusterInspector(func(c cluster.Cluster) {
|
|
mockstore.BootstrapWithSingleStore(c)
|
|
s.cluster = c
|
|
}),
|
|
)
|
|
c.Assert(err, IsNil)
|
|
|
|
s.dom, err = session.BootstrapSession(s.store)
|
|
c.Assert(err, IsNil)
|
|
}
|
|
|
|
func (s *testSerialSuite) TearDownSuite(c *C) {
|
|
if s.dom != nil {
|
|
s.dom.Close()
|
|
}
|
|
if s.store != nil {
|
|
s.store.Close()
|
|
}
|
|
}
|
|
|
|
func (s *testSerialSuite) TestChangeMaxIndexLength(c *C) {
|
|
tk := testkit.NewTestKitWithInit(c, s.store)
|
|
|
|
defer config.RestoreFunc()()
|
|
config.UpdateGlobal(func(conf *config.Config) {
|
|
conf.MaxIndexLength = config.DefMaxOfMaxIndexLength
|
|
})
|
|
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("drop table if exists t1;")
|
|
|
|
tk.MustExec("create table t (c1 varchar(3073), index(c1)) charset = ascii;")
|
|
tk.MustExec(fmt.Sprintf("create table t1 (c1 varchar(%d), index(c1)) charset = ascii;", config.DefMaxOfMaxIndexLength))
|
|
_, err := tk.Exec(fmt.Sprintf("create table t2 (c1 varchar(%d), index(c1)) charset = ascii;", config.DefMaxOfMaxIndexLength+1))
|
|
c.Assert(err.Error(), Equals, "[ddl:1071]Specified key was too long; max key length is 12288 bytes")
|
|
tk.MustExec("drop table t, t1")
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestPrimaryKey(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("drop database if exists test_primary_key;")
|
|
tk.MustExec("create database test_primary_key;")
|
|
tk.MustExec("use test_primary_key;")
|
|
tk.Se.GetSessionVars().EnableClusteredIndex = variable.ClusteredIndexDefModeIntOnly
|
|
|
|
// Test add/drop primary key on a plain table.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a int, b varchar(10));")
|
|
tk.MustGetErrCode("alter table t add primary key(a) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustExec("alter table t add primary key(a) nonclustered;")
|
|
tk.MustExec("alter table t drop primary key;")
|
|
tk.MustExec("alter table t add primary key(a) nonclustered;")
|
|
tk.MustExec("drop index `primary` on t;")
|
|
tk.MustExec("alter table t add primary key(a);") // implicit nonclustered
|
|
tk.MustExec("drop index `primary` on t;")
|
|
tk.MustGetErrCode("drop index `primary` on t;", errno.ErrCantDropFieldOrKey)
|
|
|
|
// Test add/drop primary key on a PKIsHandle table.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a int, b varchar(10), primary key(a) clustered);")
|
|
tk.MustGetErrCode("alter table t drop primary key;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(a) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(a) nonclustered;", mysql.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(a);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
tk.MustGetErrCode("alter table t add primary key(b) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(b) nonclustered;", errno.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(b);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
|
|
// Test add/drop primary key on a nonclustered primary key table.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a int, b varchar(10), primary key(a) nonclustered);")
|
|
tk.MustGetErrCode("alter table t add primary key(a) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(a) nonclustered;", errno.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(a);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
tk.MustGetErrCode("alter table t add primary key(b) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(b) nonclustered;", errno.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(b);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
tk.MustExec("alter table t drop primary key;")
|
|
|
|
// Test add/drop primary key on a CommonHandle key table.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a int, b varchar(10), primary key(b) clustered);")
|
|
tk.MustGetErrCode("alter table t drop primary key;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(a) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(a) nonclustered;", errno.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(a);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
tk.MustGetErrCode("alter table t add primary key(b) clustered;", errno.ErrUnsupportedDDLOperation)
|
|
tk.MustGetErrCode("alter table t add primary key(b) nonclustered;", errno.ErrMultiplePriKey)
|
|
tk.MustGetErrCode("alter table t add primary key(b);", errno.ErrMultiplePriKey) // implicit nonclustered
|
|
|
|
// Test add/drop primary key when the column&index name is `primary`.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (`primary` int);")
|
|
tk.MustExec("alter table t add index (`primary`);")
|
|
tk.MustGetErrCode("drop index `primary` on t;", errno.ErrCantDropFieldOrKey)
|
|
|
|
// The primary key cannot be invisible, for the case pk_is_handle.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustGetErrCode("create table t(c1 int not null, primary key(c1) invisible);", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustExec("create table t (a int, b int not null, primary key(a), unique(b) invisible);")
|
|
tk.MustExec("drop table t;")
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestDropAutoIncrementIndex(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t1")
|
|
tk.MustExec("create table t1 (a int(11) not null auto_increment key, b int(11), c bigint, unique key (a, b, c))")
|
|
tk.MustExec("alter table t1 drop index a")
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestMultiRegionGetTableEndHandle(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("drop database if exists test_get_endhandle")
|
|
tk.MustExec("create database test_get_endhandle")
|
|
tk.MustExec("use test_get_endhandle")
|
|
|
|
tk.MustExec("create table t(a bigint PRIMARY KEY, b int)")
|
|
var builder strings.Builder
|
|
fmt.Fprintf(&builder, "insert into t values ")
|
|
for i := 0; i < 1000; i++ {
|
|
fmt.Fprintf(&builder, "(%v, %v),", i, i)
|
|
}
|
|
sql := builder.String()
|
|
tk.MustExec(sql[:len(sql)-1])
|
|
|
|
// Get table ID for split.
|
|
dom := domain.GetDomain(tk.Se)
|
|
is := dom.InfoSchema()
|
|
tbl, err := is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t"))
|
|
c.Assert(err, IsNil)
|
|
tblID := tbl.Meta().ID
|
|
|
|
d := s.dom.DDL()
|
|
testCtx := newTestMaxTableRowIDContext(c, d, tbl)
|
|
|
|
// Split the table.
|
|
tableStart := tablecodec.GenTableRecordPrefix(tblID)
|
|
s.cluster.SplitKeys(tableStart, tableStart.PrefixNext(), 100)
|
|
|
|
maxHandle, emptyTable := getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, Equals, kv.IntHandle(999))
|
|
|
|
tk.MustExec("insert into t values(10000, 1000)")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, Equals, kv.IntHandle(10000))
|
|
|
|
tk.MustExec("insert into t values(-1, 1000)")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, Equals, kv.IntHandle(10000))
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestGetTableEndHandle(c *C) {
|
|
// TestGetTableEndHandle test ddl.GetTableMaxHandle method, which will return the max row id of the table.
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("drop database if exists test_get_endhandle")
|
|
tk.MustExec("create database test_get_endhandle")
|
|
tk.MustExec("use test_get_endhandle")
|
|
// Test PK is handle.
|
|
tk.MustExec("create table t(a bigint PRIMARY KEY, b int)")
|
|
|
|
is := s.dom.InfoSchema()
|
|
d := s.dom.DDL()
|
|
tbl, err := is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t"))
|
|
c.Assert(err, IsNil)
|
|
|
|
testCtx := newTestMaxTableRowIDContext(c, d, tbl)
|
|
// test empty table
|
|
checkGetMaxTableRowID(testCtx, s.store, true, nil)
|
|
|
|
tk.MustExec("insert into t values(-1, 1)")
|
|
checkGetMaxTableRowID(testCtx, s.store, false, kv.IntHandle(-1))
|
|
|
|
tk.MustExec("insert into t values(9223372036854775806, 1)")
|
|
checkGetMaxTableRowID(testCtx, s.store, false, kv.IntHandle(9223372036854775806))
|
|
|
|
tk.MustExec("insert into t values(9223372036854775807, 1)")
|
|
checkGetMaxTableRowID(testCtx, s.store, false, kv.IntHandle(9223372036854775807))
|
|
|
|
tk.MustExec("insert into t values(10, 1)")
|
|
tk.MustExec("insert into t values(102149142, 1)")
|
|
checkGetMaxTableRowID(testCtx, s.store, false, kv.IntHandle(9223372036854775807))
|
|
|
|
tk.MustExec("create table t1(a bigint PRIMARY KEY, b int)")
|
|
|
|
var builder strings.Builder
|
|
fmt.Fprintf(&builder, "insert into t1 values ")
|
|
for i := 0; i < 1000; i++ {
|
|
fmt.Fprintf(&builder, "(%v, %v),", i, i)
|
|
}
|
|
sql := builder.String()
|
|
tk.MustExec(sql[:len(sql)-1])
|
|
|
|
is = s.dom.InfoSchema()
|
|
testCtx.tbl, err = is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t1"))
|
|
c.Assert(err, IsNil)
|
|
checkGetMaxTableRowID(testCtx, s.store, false, kv.IntHandle(999))
|
|
|
|
// Test PK is not handle
|
|
tk.MustExec("create table t2(a varchar(255))")
|
|
|
|
is = s.dom.InfoSchema()
|
|
testCtx.tbl, err = is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t2"))
|
|
c.Assert(err, IsNil)
|
|
checkGetMaxTableRowID(testCtx, s.store, true, nil)
|
|
|
|
builder.Reset()
|
|
fmt.Fprintf(&builder, "insert into t2 values ")
|
|
for i := 0; i < 1000; i++ {
|
|
fmt.Fprintf(&builder, "(%v),", i)
|
|
}
|
|
sql = builder.String()
|
|
tk.MustExec(sql[:len(sql)-1])
|
|
|
|
result := tk.MustQuery("select MAX(_tidb_rowid) from t2")
|
|
maxHandle, emptyTable := getMaxTableHandle(testCtx, s.store)
|
|
result.Check(testkit.Rows(fmt.Sprintf("%v", maxHandle.IntValue())))
|
|
c.Assert(emptyTable, IsFalse)
|
|
|
|
tk.MustExec("insert into t2 values(100000)")
|
|
result = tk.MustQuery("select MAX(_tidb_rowid) from t2")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
result.Check(testkit.Rows(fmt.Sprintf("%v", maxHandle.IntValue())))
|
|
c.Assert(emptyTable, IsFalse)
|
|
|
|
tk.MustExec(fmt.Sprintf("insert into t2 values(%v)", math.MaxInt64-1))
|
|
result = tk.MustQuery("select MAX(_tidb_rowid) from t2")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
result.Check(testkit.Rows(fmt.Sprintf("%v", maxHandle.IntValue())))
|
|
c.Assert(emptyTable, IsFalse)
|
|
|
|
tk.MustExec(fmt.Sprintf("insert into t2 values(%v)", math.MaxInt64))
|
|
result = tk.MustQuery("select MAX(_tidb_rowid) from t2")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
result.Check(testkit.Rows(fmt.Sprintf("%v", maxHandle.IntValue())))
|
|
c.Assert(emptyTable, IsFalse)
|
|
|
|
tk.MustExec("insert into t2 values(100)")
|
|
result = tk.MustQuery("select MAX(_tidb_rowid) from t2")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
result.Check(testkit.Rows(fmt.Sprintf("%v", maxHandle.IntValue())))
|
|
c.Assert(emptyTable, IsFalse)
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestMultiRegionGetTableEndCommonHandle(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("drop database if exists test_get_endhandle")
|
|
tk.MustExec("create database test_get_endhandle")
|
|
tk.MustExec("use test_get_endhandle")
|
|
tk.Se.GetSessionVars().EnableClusteredIndex = variable.ClusteredIndexDefModeOn
|
|
|
|
tk.MustExec("create table t(a varchar(20), b int, c float, d bigint, primary key (a, b, c))")
|
|
var builder strings.Builder
|
|
fmt.Fprintf(&builder, "insert into t values ")
|
|
for i := 0; i < 1000; i++ {
|
|
fmt.Fprintf(&builder, "('%v', %v, %v, %v),", i, i, i, i)
|
|
}
|
|
sql := builder.String()
|
|
tk.MustExec(sql[:len(sql)-1])
|
|
|
|
// Get table ID for split.
|
|
dom := domain.GetDomain(tk.Se)
|
|
is := dom.InfoSchema()
|
|
tbl, err := is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t"))
|
|
c.Assert(err, IsNil)
|
|
tblID := tbl.Meta().ID
|
|
|
|
d := s.dom.DDL()
|
|
testCtx := newTestMaxTableRowIDContext(c, d, tbl)
|
|
|
|
// Split the table.
|
|
tableStart := tablecodec.GenTableRecordPrefix(tblID)
|
|
s.cluster.SplitKeys(tableStart, tableStart.PrefixNext(), 100)
|
|
|
|
maxHandle, emptyTable := getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, HandleEquals, MustNewCommonHandle(c, "999", 999, 999))
|
|
|
|
tk.MustExec("insert into t values('a', 1, 1, 1)")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, HandleEquals, MustNewCommonHandle(c, "a", 1, 1))
|
|
|
|
tk.MustExec("insert into t values('0000', 1, 1, 1)")
|
|
maxHandle, emptyTable = getMaxTableHandle(testCtx, s.store)
|
|
c.Assert(emptyTable, IsFalse)
|
|
c.Assert(maxHandle, HandleEquals, MustNewCommonHandle(c, "a", 1, 1))
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestGetTableEndCommonHandle(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("drop database if exists test_get_endhandle")
|
|
tk.MustExec("create database test_get_endhandle")
|
|
tk.MustExec("use test_get_endhandle")
|
|
tk.Se.GetSessionVars().EnableClusteredIndex = variable.ClusteredIndexDefModeOn
|
|
|
|
tk.MustExec("create table t(a varchar(15), b bigint, c int, primary key (a, b))")
|
|
tk.MustExec("create table t1(a varchar(15), b bigint, c int, primary key (a(2), b))")
|
|
|
|
is := s.dom.InfoSchema()
|
|
d := s.dom.DDL()
|
|
tbl, err := is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t"))
|
|
c.Assert(err, IsNil)
|
|
testCtx := newTestMaxTableRowIDContext(c, d, tbl)
|
|
|
|
// test empty table
|
|
checkGetMaxTableRowID(testCtx, s.store, true, nil)
|
|
tk.MustExec("insert into t values('abc', 1, 10)")
|
|
expectedHandle := MustNewCommonHandle(c, "abc", 1)
|
|
checkGetMaxTableRowID(testCtx, s.store, false, expectedHandle)
|
|
tk.MustExec("insert into t values('abchzzzzzzzz', 1, 10)")
|
|
expectedHandle = MustNewCommonHandle(c, "abchzzzzzzzz", 1)
|
|
checkGetMaxTableRowID(testCtx, s.store, false, expectedHandle)
|
|
tk.MustExec("insert into t values('a', 1, 10)")
|
|
tk.MustExec("insert into t values('ab', 1, 10)")
|
|
checkGetMaxTableRowID(testCtx, s.store, false, expectedHandle)
|
|
|
|
// Test MaxTableRowID with prefixed primary key.
|
|
tbl, err = is.TableByName(model.NewCIStr("test_get_endhandle"), model.NewCIStr("t1"))
|
|
c.Assert(err, IsNil)
|
|
d = s.dom.DDL()
|
|
testCtx = newTestMaxTableRowIDContext(c, d, tbl)
|
|
checkGetMaxTableRowID(testCtx, s.store, true, nil)
|
|
tk.MustExec("insert into t1 values('abccccc', 1, 10)")
|
|
expectedHandle = MustNewCommonHandle(c, "ab", 1)
|
|
checkGetMaxTableRowID(testCtx, s.store, false, expectedHandle)
|
|
tk.MustExec("insert into t1 values('azzzz', 1, 10)")
|
|
expectedHandle = MustNewCommonHandle(c, "az", 1)
|
|
checkGetMaxTableRowID(testCtx, s.store, false, expectedHandle)
|
|
}
|
|
|
|
func (s *testSerialSuite) TestCreateTableWithLike(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
// for the same database
|
|
tk.MustExec("create database ctwl_db")
|
|
tk.MustExec("use ctwl_db")
|
|
tk.MustExec("create table tt(id int primary key)")
|
|
tk.MustExec("create table t (c1 int not null auto_increment, c2 int, constraint cc foreign key (c2) references tt(id), primary key(c1)) auto_increment = 10")
|
|
tk.MustExec("insert into t set c2=1")
|
|
tk.MustExec("create table t1 like ctwl_db.t")
|
|
tk.MustExec("insert into t1 set c2=11")
|
|
tk.MustExec("create table t2 (like ctwl_db.t1)")
|
|
tk.MustExec("insert into t2 set c2=12")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("10 1"))
|
|
tk.MustQuery("select * from t1").Check(testkit.Rows("1 11"))
|
|
tk.MustQuery("select * from t2").Check(testkit.Rows("1 12"))
|
|
ctx := tk.Se.(sessionctx.Context)
|
|
is := domain.GetDomain(ctx).InfoSchema()
|
|
tbl1, err := is.TableByName(model.NewCIStr("ctwl_db"), model.NewCIStr("t1"))
|
|
c.Assert(err, IsNil)
|
|
tbl1Info := tbl1.Meta()
|
|
c.Assert(tbl1Info.ForeignKeys, IsNil)
|
|
c.Assert(tbl1Info.PKIsHandle, Equals, true)
|
|
col := tbl1Info.Columns[0]
|
|
hasNotNull := mysql.HasNotNullFlag(col.Flag)
|
|
c.Assert(hasNotNull, IsTrue)
|
|
tbl2, err := is.TableByName(model.NewCIStr("ctwl_db"), model.NewCIStr("t2"))
|
|
c.Assert(err, IsNil)
|
|
tbl2Info := tbl2.Meta()
|
|
c.Assert(tbl2Info.ForeignKeys, IsNil)
|
|
c.Assert(tbl2Info.PKIsHandle, Equals, true)
|
|
c.Assert(mysql.HasNotNullFlag(tbl2Info.Columns[0].Flag), IsTrue)
|
|
|
|
// for different databases
|
|
tk.MustExec("create database ctwl_db1")
|
|
tk.MustExec("use ctwl_db1")
|
|
tk.MustExec("create table t1 like ctwl_db.t")
|
|
tk.MustExec("insert into t1 set c2=11")
|
|
tk.MustQuery("select * from t1").Check(testkit.Rows("1 11"))
|
|
is = domain.GetDomain(ctx).InfoSchema()
|
|
tbl1, err = is.TableByName(model.NewCIStr("ctwl_db1"), model.NewCIStr("t1"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl1.Meta().ForeignKeys, IsNil)
|
|
|
|
// for table partition
|
|
tk.MustExec("use ctwl_db")
|
|
tk.MustExec("create table pt1 (id int) partition by range columns (id) (partition p0 values less than (10))")
|
|
tk.MustExec("insert into pt1 values (1),(2),(3),(4);")
|
|
tk.MustExec("create table ctwl_db1.pt1 like ctwl_db.pt1;")
|
|
tk.MustQuery("select * from ctwl_db1.pt1").Check(testkit.Rows())
|
|
|
|
// Test create table like for partition table.
|
|
atomic.StoreUint32(&ddl.EnableSplitTableRegion, 1)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("set @@global.tidb_scatter_region=1;")
|
|
tk.MustExec("drop table if exists partition_t;")
|
|
tk.MustExec("create table partition_t (a int, b int,index(a)) partition by hash (a) partitions 3")
|
|
tk.MustExec("drop table if exists t1;")
|
|
tk.MustExec("create table t1 like partition_t")
|
|
re := tk.MustQuery("show table t1 regions")
|
|
rows := re.Rows()
|
|
c.Assert(len(rows), Equals, 3)
|
|
tbl := testGetTableByName(c, tk.Se, "test", "t1")
|
|
partitionDef := tbl.Meta().GetPartitionInfo().Definitions
|
|
c.Assert(rows[0][1], Matches, fmt.Sprintf("t_%d_.*", partitionDef[0].ID))
|
|
c.Assert(rows[1][1], Matches, fmt.Sprintf("t_%d_.*", partitionDef[1].ID))
|
|
c.Assert(rows[2][1], Matches, fmt.Sprintf("t_%d_.*", partitionDef[2].ID))
|
|
|
|
// Test pre-split table region when create table like.
|
|
tk.MustExec("drop table if exists t_pre")
|
|
tk.MustExec("create table t_pre (a int, b int) shard_row_id_bits = 2 pre_split_regions=2;")
|
|
tk.MustExec("drop table if exists t2;")
|
|
tk.MustExec("create table t2 like t_pre")
|
|
re = tk.MustQuery("show table t2 regions")
|
|
rows = re.Rows()
|
|
// Table t2 which create like t_pre should have 4 regions now.
|
|
c.Assert(len(rows), Equals, 4)
|
|
tbl = testGetTableByName(c, tk.Se, "test", "t2")
|
|
c.Assert(rows[1][1], Equals, fmt.Sprintf("t_%d_r_2305843009213693952", tbl.Meta().ID))
|
|
c.Assert(rows[2][1], Equals, fmt.Sprintf("t_%d_r_4611686018427387904", tbl.Meta().ID))
|
|
c.Assert(rows[3][1], Equals, fmt.Sprintf("t_%d_r_6917529027641081856", tbl.Meta().ID))
|
|
// Test after truncate table the region is also splited.
|
|
tk.MustExec("truncate table t2")
|
|
re = tk.MustQuery("show table t2 regions")
|
|
rows = re.Rows()
|
|
c.Assert(len(rows), Equals, 4)
|
|
tbl = testGetTableByName(c, tk.Se, "test", "t2")
|
|
c.Assert(rows[1][1], Equals, fmt.Sprintf("t_%d_r_2305843009213693952", tbl.Meta().ID))
|
|
c.Assert(rows[2][1], Equals, fmt.Sprintf("t_%d_r_4611686018427387904", tbl.Meta().ID))
|
|
c.Assert(rows[3][1], Equals, fmt.Sprintf("t_%d_r_6917529027641081856", tbl.Meta().ID))
|
|
|
|
defer atomic.StoreUint32(&ddl.EnableSplitTableRegion, 0)
|
|
|
|
// for failure table cases
|
|
tk.MustExec("use ctwl_db")
|
|
failSQL := "create table t1 like test_not_exist.t"
|
|
tk.MustGetErrCode(failSQL, mysql.ErrNoSuchTable)
|
|
failSQL = "create table t1 like test.t_not_exist"
|
|
tk.MustGetErrCode(failSQL, mysql.ErrNoSuchTable)
|
|
failSQL = "create table t1 (like test_not_exist.t)"
|
|
tk.MustGetErrCode(failSQL, mysql.ErrNoSuchTable)
|
|
failSQL = "create table test_not_exis.t1 like ctwl_db.t"
|
|
tk.MustGetErrCode(failSQL, mysql.ErrBadDB)
|
|
failSQL = "create table t1 like ctwl_db.t"
|
|
tk.MustGetErrCode(failSQL, mysql.ErrTableExists)
|
|
|
|
// test failure for wrong object cases
|
|
tk.MustExec("drop view if exists v")
|
|
tk.MustExec("create view v as select 1 from dual")
|
|
tk.MustGetErrCode("create table viewTable like v", mysql.ErrWrongObject)
|
|
tk.MustExec("drop sequence if exists seq")
|
|
tk.MustExec("create sequence seq")
|
|
tk.MustGetErrCode("create table sequenceTable like seq", mysql.ErrWrongObject)
|
|
|
|
tk.MustExec("drop database ctwl_db")
|
|
tk.MustExec("drop database ctwl_db1")
|
|
}
|
|
|
|
func (s *testSerialSuite) TestCreateTableWithLikeAtTemporaryMode(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
|
|
// Test create table like at temporary mode.
|
|
tk.MustExec("set tidb_enable_global_temporary_table=true")
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists temporary_table;")
|
|
tk.MustExec("create global temporary table temporary_table (a int, b int,index(a)) on commit delete rows")
|
|
tk.MustExec("drop table if exists temporary_table_t1;")
|
|
_, err := tk.Exec("create table temporary_table_t1 like temporary_table")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("create table like").Error())
|
|
tk.MustExec("drop table if exists temporary_table;")
|
|
|
|
// Test create temporary table like.
|
|
// Test auto_random.
|
|
tk.MustExec("drop table if exists auto_random_table")
|
|
_, err = tk.Exec("create table auto_random_table (a bigint primary key auto_random(3), b varchar(255));")
|
|
defer tk.MustExec("drop table if exists auto_random_table")
|
|
tk.MustExec("drop table if exists auto_random_temporary_global")
|
|
_, err = tk.Exec("create global temporary table auto_random_temporary_global like auto_random_table on commit delete rows;")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("auto_random").Error())
|
|
|
|
// Test pre split regions.
|
|
tk.MustExec("drop table if exists table_pre_split")
|
|
_, err = tk.Exec("create table table_pre_split(id int) shard_row_id_bits = 2 pre_split_regions=2;")
|
|
defer tk.MustExec("drop table if exists table_pre_split")
|
|
tk.MustExec("drop table if exists temporary_table_pre_split")
|
|
_, err = tk.Exec("create global temporary table temporary_table_pre_split like table_pre_split ON COMMIT DELETE ROWS;")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("pre split regions").Error())
|
|
|
|
// Test shard_row_id_bits.
|
|
tk.MustExec("drop table if exists shard_row_id_table, shard_row_id_temporary_table, shard_row_id_table_plus, shard_row_id_temporary_table_plus")
|
|
_, err = tk.Exec("create table shard_row_id_table (a int) shard_row_id_bits = 5;")
|
|
_, err = tk.Exec("create global temporary table shard_row_id_temporary_table like shard_row_id_table on commit delete rows;")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("shard_row_id_bits").Error())
|
|
tk.MustExec("create table shard_row_id_table_plus (a int);")
|
|
tk.MustExec("create global temporary table shard_row_id_temporary_table_plus (a int) on commit delete rows;")
|
|
defer tk.MustExec("drop table if exists shard_row_id_table, shard_row_id_temporary_table, shard_row_id_table_plus, shard_row_id_temporary_table_plus")
|
|
_, err = tk.Exec("alter table shard_row_id_temporary_table_plus shard_row_id_bits = 4;")
|
|
c.Assert(err.Error(), Equals, ddl.ErrOptOnTemporaryTable.GenWithStackByArgs("shard_row_id_bits").Error())
|
|
|
|
// Test partition.
|
|
tk.MustExec("drop table if exists global_partition_table;")
|
|
tk.MustExec("create table global_partition_table (a int, b int) partition by hash(a) partitions 3;")
|
|
defer tk.MustExec("drop table if exists global_partition_table;")
|
|
tk.MustGetErrCode("create global temporary table global_partition_temp_table like global_partition_table ON COMMIT DELETE ROWS;",
|
|
errno.ErrPartitionNoTemporary)
|
|
// Test virtual columns.
|
|
tk.MustExec("drop table if exists test_gv_ddl, test_gv_ddl_temp")
|
|
tk.MustExec(`create table test_gv_ddl(a int, b int as (a+8) virtual, c int as (b + 2) stored)`)
|
|
tk.MustExec(`create global temporary table test_gv_ddl_temp like test_gv_ddl on commit delete rows;`)
|
|
defer tk.MustExec("drop table if exists test_gv_ddl_temp, test_gv_ddl")
|
|
is := tk.Se.(sessionctx.Context).GetInfoSchema().(infoschema.InfoSchema)
|
|
table, err := is.TableByName(model.NewCIStr("test"), model.NewCIStr("test_gv_ddl"))
|
|
c.Assert(err, IsNil)
|
|
testCases := []struct {
|
|
generatedExprString string
|
|
generatedStored bool
|
|
}{
|
|
{"", false},
|
|
{"`a` + 8", false},
|
|
{"`b` + 2", true},
|
|
}
|
|
for i, column := range table.Meta().Columns {
|
|
c.Assert(column.GeneratedExprString, Equals, testCases[i].generatedExprString)
|
|
c.Assert(column.GeneratedStored, Equals, testCases[i].generatedStored)
|
|
}
|
|
result := tk.MustQuery(`DESC test_gv_ddl_temp`)
|
|
result.Check(testkit.Rows(`a int(11) YES <nil> `, `b int(11) YES <nil> VIRTUAL GENERATED`, `c int(11) YES <nil> STORED GENERATED`))
|
|
tk.MustExec("begin;")
|
|
tk.MustExec("insert into test_gv_ddl_temp values (1, default, default)")
|
|
tk.MustQuery("select * from test_gv_ddl_temp").Check(testkit.Rows("1 9 11"))
|
|
_, err = tk.Exec("commit")
|
|
c.Assert(err, IsNil)
|
|
|
|
// Test foreign key.
|
|
tk.MustExec("drop table if exists test_foreign_key, t1")
|
|
tk.MustExec("create table t1 (a int, b int);")
|
|
tk.MustExec("create table test_foreign_key (c int,d int,foreign key (d) references t1 (b));")
|
|
defer tk.MustExec("drop table if exists test_foreign_key, t1;")
|
|
tk.MustExec("create global temporary table test_foreign_key_temp like test_foreign_key on commit delete rows;")
|
|
is = tk.Se.(sessionctx.Context).GetInfoSchema().(infoschema.InfoSchema)
|
|
table, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("test_foreign_key_temp"))
|
|
c.Assert(err, IsNil)
|
|
tableInfo := table.Meta()
|
|
c.Assert(len(tableInfo.ForeignKeys), Equals, 0)
|
|
}
|
|
|
|
// TestCancelAddIndex1 tests canceling ddl job when the add index worker is not started.
|
|
func (s *testSerialSuite) TestCancelAddIndexPanic(c *C) {
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/errorMockPanic", `return(true)`), IsNil)
|
|
defer func() {
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/errorMockPanic"), IsNil)
|
|
}()
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t")
|
|
tk.MustExec("create table t(c1 int, c2 int)")
|
|
defer tk.MustExec("drop table t;")
|
|
for i := 0; i < 5; i++ {
|
|
tk.MustExec("insert into t values (?, ?)", i, i)
|
|
}
|
|
var checkErr error
|
|
oldReorgWaitTimeout := ddl.ReorgWaitTimeout
|
|
ddl.ReorgWaitTimeout = 50 * time.Millisecond
|
|
defer func() { ddl.ReorgWaitTimeout = oldReorgWaitTimeout }()
|
|
hook := &ddl.TestDDLCallback{}
|
|
hook.OnJobRunBeforeExported = func(job *model.Job) {
|
|
if job.Type == model.ActionAddIndex && job.State == model.JobStateRunning && job.SchemaState == model.StateWriteReorganization && job.SnapshotVer != 0 {
|
|
jobIDs := []int64{job.ID}
|
|
hookCtx := mock.NewContext()
|
|
hookCtx.Store = s.store
|
|
err := hookCtx.NewTxn(context.Background())
|
|
if err != nil {
|
|
checkErr = errors.Trace(err)
|
|
return
|
|
}
|
|
txn, err := hookCtx.Txn(true)
|
|
if err != nil {
|
|
checkErr = errors.Trace(err)
|
|
return
|
|
}
|
|
errs, err := admin.CancelJobs(txn, jobIDs)
|
|
if err != nil {
|
|
checkErr = errors.Trace(err)
|
|
return
|
|
}
|
|
if errs[0] != nil {
|
|
checkErr = errors.Trace(errs[0])
|
|
return
|
|
}
|
|
txn, err = hookCtx.Txn(true)
|
|
if err != nil {
|
|
checkErr = errors.Trace(err)
|
|
return
|
|
}
|
|
checkErr = txn.Commit(context.Background())
|
|
}
|
|
}
|
|
origHook := s.dom.DDL().GetHook()
|
|
defer s.dom.DDL().(ddl.DDLForTest).SetHook(origHook)
|
|
s.dom.DDL().(ddl.DDLForTest).SetHook(hook)
|
|
rs, err := tk.Exec("alter table t add index idx_c2(c2)")
|
|
if rs != nil {
|
|
rs.Close()
|
|
}
|
|
c.Assert(checkErr, IsNil)
|
|
c.Assert(err, NotNil)
|
|
errMsg := err.Error()
|
|
// Cancelling the job can either succeed or not, it depends on whether the cancelled job takes affect.
|
|
// For now, there's no way to guarantee that cancelling will always take effect.
|
|
// TODO: After issue #17904 is fixed, there is no need to tolerate it here.
|
|
c.Assert(strings.HasPrefix(errMsg, "[ddl:8214]Cancelled DDL job") || strings.HasPrefix(errMsg, "[ddl:8211]DDL job rollback"), IsTrue)
|
|
}
|
|
|
|
func (s *testSerialSuite) TestRecoverTableByJobID(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists test_recover")
|
|
tk.MustExec("use test_recover")
|
|
tk.MustExec("drop table if exists t_recover")
|
|
tk.MustExec("create table t_recover (a int);")
|
|
defer func(originGC bool) {
|
|
if originGC {
|
|
ddl.EmulatorGCEnable()
|
|
} else {
|
|
ddl.EmulatorGCDisable()
|
|
}
|
|
}(ddl.IsEmulatorGCEnable())
|
|
|
|
// disable emulator GC.
|
|
// Otherwise emulator GC will delete table record as soon as possible after execute drop table ddl.
|
|
ddl.EmulatorGCDisable()
|
|
gcTimeFormat := "20060102-15:04:05 -0700 MST"
|
|
timeBeforeDrop := time.Now().Add(0 - 48*60*60*time.Second).Format(gcTimeFormat)
|
|
timeAfterDrop := time.Now().Add(48 * 60 * 60 * time.Second).Format(gcTimeFormat)
|
|
safePointSQL := `INSERT HIGH_PRIORITY INTO mysql.tidb VALUES ('tikv_gc_safe_point', '%[1]s', '')
|
|
ON DUPLICATE KEY
|
|
UPDATE variable_value = '%[1]s'`
|
|
// clear GC variables first.
|
|
tk.MustExec("delete from mysql.tidb where variable_name in ( 'tikv_gc_safe_point','tikv_gc_enable' )")
|
|
|
|
tk.MustExec("insert into t_recover values (1),(2),(3)")
|
|
tk.MustExec("drop table t_recover")
|
|
|
|
getDDLJobID := func(table, tp string) int64 {
|
|
rs, err := tk.Exec("admin show ddl jobs")
|
|
c.Assert(err, IsNil)
|
|
rows, err := session.GetRows4Test(context.Background(), tk.Se, rs)
|
|
c.Assert(err, IsNil)
|
|
for _, row := range rows {
|
|
if row.GetString(1) == table && row.GetString(3) == tp {
|
|
return row.GetInt64(0)
|
|
}
|
|
}
|
|
c.Errorf("can't find %s table of %s", tp, table)
|
|
return -1
|
|
}
|
|
jobID := getDDLJobID("test_recover", "drop table")
|
|
|
|
// if GC safe point is not exists in mysql.tidb
|
|
_, err := tk.Exec(fmt.Sprintf("recover table by job %d", jobID))
|
|
c.Assert(err, NotNil)
|
|
c.Assert(err.Error(), Equals, "can not get 'tikv_gc_safe_point'")
|
|
// set GC safe point
|
|
tk.MustExec(fmt.Sprintf(safePointSQL, timeBeforeDrop))
|
|
|
|
// if GC enable is not exists in mysql.tidb
|
|
tk.MustExec(fmt.Sprintf("recover table by job %d", jobID))
|
|
tk.MustExec("DROP TABLE t_recover")
|
|
|
|
err = gcutil.EnableGC(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
|
|
// recover job is before GC safe point
|
|
tk.MustExec(fmt.Sprintf(safePointSQL, timeAfterDrop))
|
|
_, err = tk.Exec(fmt.Sprintf("recover table by job %d", jobID))
|
|
c.Assert(err, NotNil)
|
|
c.Assert(strings.Contains(err.Error(), "snapshot is older than GC safe point"), Equals, true)
|
|
|
|
// set GC safe point
|
|
tk.MustExec(fmt.Sprintf(safePointSQL, timeBeforeDrop))
|
|
// if there is a new table with the same name, should return failed.
|
|
tk.MustExec("create table t_recover (a int);")
|
|
_, err = tk.Exec(fmt.Sprintf("recover table by job %d", jobID))
|
|
c.Assert(err.Error(), Equals, infoschema.ErrTableExists.GenWithStackByArgs("t_recover").Error())
|
|
|
|
// drop the new table with the same name, then recover table.
|
|
tk.MustExec("drop table t_recover")
|
|
|
|
// do recover table.
|
|
tk.MustExec(fmt.Sprintf("recover table by job %d", jobID))
|
|
|
|
// check recover table meta and data record.
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3"))
|
|
// check recover table autoID.
|
|
tk.MustExec("insert into t_recover values (4),(5),(6)")
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3", "4", "5", "6"))
|
|
|
|
// recover table by none exits job.
|
|
_, err = tk.Exec(fmt.Sprintf("recover table by job %d", 10000000))
|
|
c.Assert(err, NotNil)
|
|
|
|
// Disable GC by manual first, then after recover table, the GC enable status should also be disabled.
|
|
err = gcutil.DisableGC(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
|
|
tk.MustExec("delete from t_recover where a > 1")
|
|
tk.MustExec("drop table t_recover")
|
|
jobID = getDDLJobID("test_recover", "drop table")
|
|
|
|
tk.MustExec(fmt.Sprintf("recover table by job %d", jobID))
|
|
|
|
// check recover table meta and data record.
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1"))
|
|
// check recover table autoID.
|
|
tk.MustExec("insert into t_recover values (7),(8),(9)")
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "7", "8", "9"))
|
|
|
|
// Test for recover truncate table.
|
|
tk.MustExec("truncate table t_recover")
|
|
tk.MustExec("rename table t_recover to t_recover_new")
|
|
jobID = getDDLJobID("test_recover", "truncate table")
|
|
tk.MustExec(fmt.Sprintf("recover table by job %d", jobID))
|
|
tk.MustExec("insert into t_recover values (10)")
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "7", "8", "9", "10"))
|
|
|
|
gcEnable, err := gcutil.CheckGCEnable(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
c.Assert(gcEnable, Equals, false)
|
|
}
|
|
|
|
func (s *testSerialSuite) TestRecoverTableByJobIDFail(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists test_recover")
|
|
tk.MustExec("use test_recover")
|
|
tk.MustExec("drop table if exists t_recover")
|
|
tk.MustExec("create table t_recover (a int);")
|
|
defer func(originGC bool) {
|
|
if originGC {
|
|
ddl.EmulatorGCEnable()
|
|
} else {
|
|
ddl.EmulatorGCDisable()
|
|
}
|
|
}(ddl.IsEmulatorGCEnable())
|
|
|
|
// disable emulator GC.
|
|
// Otherwise emulator GC will delete table record as soon as possible after execute drop table ddl.
|
|
ddl.EmulatorGCDisable()
|
|
gcTimeFormat := "20060102-15:04:05 -0700 MST"
|
|
timeBeforeDrop := time.Now().Add(0 - 48*60*60*time.Second).Format(gcTimeFormat)
|
|
safePointSQL := `INSERT HIGH_PRIORITY INTO mysql.tidb VALUES ('tikv_gc_safe_point', '%[1]s', '')
|
|
ON DUPLICATE KEY
|
|
UPDATE variable_value = '%[1]s'`
|
|
|
|
tk.MustExec("insert into t_recover values (1),(2),(3)")
|
|
tk.MustExec("drop table t_recover")
|
|
|
|
rs, err := tk.Exec("admin show ddl jobs")
|
|
c.Assert(err, IsNil)
|
|
rows, err := session.GetRows4Test(context.Background(), tk.Se, rs)
|
|
c.Assert(err, IsNil)
|
|
row := rows[0]
|
|
c.Assert(row.GetString(1), Equals, "test_recover")
|
|
c.Assert(row.GetString(3), Equals, "drop table")
|
|
jobID := row.GetInt64(0)
|
|
|
|
// enableGC first
|
|
err = gcutil.EnableGC(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
tk.MustExec(fmt.Sprintf(safePointSQL, timeBeforeDrop))
|
|
|
|
// set hook
|
|
hook := &ddl.TestDDLCallback{}
|
|
hook.OnJobRunBeforeExported = func(job *model.Job) {
|
|
if job.Type == model.ActionRecoverTable {
|
|
c.Assert(failpoint.Enable("tikvclient/mockCommitError", `return(true)`), IsNil)
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/mockRecoverTableCommitErr", `return(true)`), IsNil)
|
|
}
|
|
}
|
|
origHook := s.dom.DDL().GetHook()
|
|
defer s.dom.DDL().(ddl.DDLForTest).SetHook(origHook)
|
|
s.dom.DDL().(ddl.DDLForTest).SetHook(hook)
|
|
|
|
// do recover table.
|
|
tk.MustExec(fmt.Sprintf("recover table by job %d", jobID))
|
|
c.Assert(failpoint.Disable("tikvclient/mockCommitError"), IsNil)
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/mockRecoverTableCommitErr"), IsNil)
|
|
|
|
// make sure enable GC after recover table.
|
|
enable, err := gcutil.CheckGCEnable(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
c.Assert(enable, Equals, true)
|
|
|
|
// check recover table meta and data record.
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3"))
|
|
// check recover table autoID.
|
|
tk.MustExec("insert into t_recover values (4),(5),(6)")
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3", "4", "5", "6"))
|
|
}
|
|
|
|
func (s *testSerialSuite) TestRecoverTableByTableNameFail(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists test_recover")
|
|
tk.MustExec("use test_recover")
|
|
tk.MustExec("drop table if exists t_recover")
|
|
tk.MustExec("create table t_recover (a int);")
|
|
defer func(originGC bool) {
|
|
if originGC {
|
|
ddl.EmulatorGCEnable()
|
|
} else {
|
|
ddl.EmulatorGCDisable()
|
|
}
|
|
}(ddl.IsEmulatorGCEnable())
|
|
|
|
// disable emulator GC.
|
|
// Otherwise emulator GC will delete table record as soon as possible after execute drop table ddl.
|
|
ddl.EmulatorGCDisable()
|
|
gcTimeFormat := "20060102-15:04:05 -0700 MST"
|
|
timeBeforeDrop := time.Now().Add(0 - 48*60*60*time.Second).Format(gcTimeFormat)
|
|
safePointSQL := `INSERT HIGH_PRIORITY INTO mysql.tidb VALUES ('tikv_gc_safe_point', '%[1]s', '')
|
|
ON DUPLICATE KEY
|
|
UPDATE variable_value = '%[1]s'`
|
|
|
|
tk.MustExec("insert into t_recover values (1),(2),(3)")
|
|
tk.MustExec("drop table t_recover")
|
|
|
|
// enableGC first
|
|
err := gcutil.EnableGC(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
tk.MustExec(fmt.Sprintf(safePointSQL, timeBeforeDrop))
|
|
|
|
// set hook
|
|
hook := &ddl.TestDDLCallback{}
|
|
hook.OnJobRunBeforeExported = func(job *model.Job) {
|
|
if job.Type == model.ActionRecoverTable {
|
|
c.Assert(failpoint.Enable("tikvclient/mockCommitError", `return(true)`), IsNil)
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/mockRecoverTableCommitErr", `return(true)`), IsNil)
|
|
}
|
|
}
|
|
origHook := s.dom.DDL().GetHook()
|
|
defer s.dom.DDL().(ddl.DDLForTest).SetHook(origHook)
|
|
s.dom.DDL().(ddl.DDLForTest).SetHook(hook)
|
|
|
|
// do recover table.
|
|
tk.MustExec("recover table t_recover")
|
|
c.Assert(failpoint.Disable("tikvclient/mockCommitError"), IsNil)
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/mockRecoverTableCommitErr"), IsNil)
|
|
|
|
// make sure enable GC after recover table.
|
|
enable, err := gcutil.CheckGCEnable(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
c.Assert(enable, Equals, true)
|
|
|
|
// check recover table meta and data record.
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3"))
|
|
// check recover table autoID.
|
|
tk.MustExec("insert into t_recover values (4),(5),(6)")
|
|
tk.MustQuery("select * from t_recover;").Check(testkit.Rows("1", "2", "3", "4", "5", "6"))
|
|
}
|
|
|
|
func (s *testSerialSuite) TestCancelJobByErrorCountLimit(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/mockExceedErrorLimit", `return(true)`), IsNil)
|
|
defer func() {
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/mockExceedErrorLimit"), IsNil)
|
|
}()
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t")
|
|
|
|
limit := variable.GetDDLErrorCountLimit()
|
|
tk.MustExec("set @@global.tidb_ddl_error_count_limit = 16")
|
|
err := ddlutil.LoadDDLVars(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
defer tk.MustExec(fmt.Sprintf("set @@global.tidb_ddl_error_count_limit = %d", limit))
|
|
|
|
_, err = tk.Exec("create table t (a int)")
|
|
c.Assert(err, NotNil)
|
|
c.Assert(err.Error(), Equals, "[ddl:-1]DDL job rollback, error msg: mock do job error")
|
|
}
|
|
|
|
func (s *testSerialSuite) TestTruncateTableUpdateSchemaVersionErr(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/mockTruncateTableUpdateVersionError", `return(true)`), IsNil)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t")
|
|
|
|
limit := variable.GetDDLErrorCountLimit()
|
|
tk.MustExec("set @@global.tidb_ddl_error_count_limit = 5")
|
|
err := ddlutil.LoadDDLVars(tk.Se)
|
|
c.Assert(err, IsNil)
|
|
defer tk.MustExec(fmt.Sprintf("set @@global.tidb_ddl_error_count_limit = %d", limit))
|
|
|
|
tk.MustExec("create table t (a int)")
|
|
_, err = tk.Exec("truncate table t")
|
|
c.Assert(err, NotNil)
|
|
c.Assert(err.Error(), Equals, "[ddl:-1]DDL job rollback, error msg: mock update version error")
|
|
// Disable fail point.
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/mockTruncateTableUpdateVersionError"), IsNil)
|
|
tk.MustExec("truncate table t")
|
|
}
|
|
|
|
func (s *testSerialSuite) TestCanceledJobTakeTime(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("create table t_cjtt(a int)")
|
|
|
|
hook := &ddl.TestDDLCallback{}
|
|
once := sync.Once{}
|
|
hook.OnJobUpdatedExported = func(job *model.Job) {
|
|
once.Do(func() {
|
|
err := kv.RunInNewTxn(context.Background(), s.store, false, func(ctx context.Context, txn kv.Transaction) error {
|
|
t := meta.NewMeta(txn)
|
|
return t.DropTableOrView(job.SchemaID, job.TableID, true)
|
|
})
|
|
c.Assert(err, IsNil)
|
|
})
|
|
}
|
|
origHook := s.dom.DDL().GetHook()
|
|
s.dom.DDL().(ddl.DDLForTest).SetHook(hook)
|
|
defer s.dom.DDL().(ddl.DDLForTest).SetHook(origHook)
|
|
|
|
originalWT := ddl.GetWaitTimeWhenErrorOccurred()
|
|
ddl.SetWaitTimeWhenErrorOccurred(1 * time.Second)
|
|
defer func() { ddl.SetWaitTimeWhenErrorOccurred(originalWT) }()
|
|
startTime := time.Now()
|
|
tk.MustGetErrCode("alter table t_cjtt add column b int", mysql.ErrNoSuchTable)
|
|
sub := time.Since(startTime)
|
|
c.Assert(sub, Less, ddl.GetWaitTimeWhenErrorOccurred())
|
|
}
|
|
|
|
func (s *testSerialSuite) TestTableLocksEnable(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t1")
|
|
defer tk.MustExec("drop table if exists t1")
|
|
tk.MustExec("create table t1 (a int)")
|
|
|
|
// Test for enable table lock config.
|
|
defer config.RestoreFunc()()
|
|
config.UpdateGlobal(func(conf *config.Config) {
|
|
conf.EnableTableLock = false
|
|
})
|
|
|
|
tk.MustExec("lock tables t1 write")
|
|
checkTableLock(c, tk.Se, "test", "t1", model.TableLockNone)
|
|
}
|
|
|
|
func (s *testSerialDBSuite) TestAutoRandomOnTemporaryTable(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists auto_random_temporary")
|
|
tk.MustExec("set tidb_enable_global_temporary_table=true")
|
|
_, err := tk.Exec("create global temporary table auto_random_temporary (a bigint primary key auto_random(3), b varchar(255)) on commit delete rows;")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("auto_random").Error())
|
|
tk.MustExec("set @@tidb_enable_noop_functions = 1")
|
|
_, err = tk.Exec("create temporary table t(a bigint key auto_random);")
|
|
c.Assert(err.Error(), Equals, core.ErrOptOnTemporaryTable.GenWithStackByArgs("auto_random").Error())
|
|
}
|
|
|
|
func (s *testSerialDBSuite) TestAutoRandom(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists auto_random_db")
|
|
defer tk.MustExec("drop database if exists auto_random_db")
|
|
tk.MustExec("use auto_random_db")
|
|
databaseName, tableName := "auto_random_db", "t"
|
|
tk.MustExec("drop table if exists t")
|
|
tk.MustExec("set @@allow_auto_random_explicit_insert = true")
|
|
|
|
assertInvalidAutoRandomErr := func(sql string, errMsg string, args ...interface{}) {
|
|
_, err := tk.Exec(sql)
|
|
c.Assert(err, NotNil)
|
|
c.Assert(err.Error(), Equals, ddl.ErrInvalidAutoRandom.GenWithStackByArgs(fmt.Sprintf(errMsg, args...)).Error())
|
|
}
|
|
|
|
assertPKIsNotHandle := func(sql, errCol string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomPKisNotHandleErrMsg, errCol)
|
|
}
|
|
assertAlterValue := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomAlterErrMsg)
|
|
}
|
|
assertOnlyChangeFromAutoIncPK := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomAlterChangeFromAutoInc)
|
|
}
|
|
assertDecreaseBitErr := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomDecreaseBitErrMsg)
|
|
}
|
|
assertWithAutoInc := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomIncompatibleWithAutoIncErrMsg)
|
|
}
|
|
assertOverflow := func(sql, colName string, maxAutoRandBits, actualAutoRandBits uint64) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomOverflowErrMsg, maxAutoRandBits, actualAutoRandBits, colName)
|
|
}
|
|
assertMaxOverflow := func(sql, colName string, autoRandBits uint64) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomOverflowErrMsg, autoid.MaxAutoRandomBits, autoRandBits, colName)
|
|
}
|
|
assertModifyColType := func(sql string) {
|
|
tk.MustGetErrCode(sql, errno.ErrUnsupportedDDLOperation)
|
|
}
|
|
assertDefault := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomIncompatibleWithDefaultValueErrMsg)
|
|
}
|
|
assertNonPositive := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomNonPositive)
|
|
}
|
|
assertBigIntOnly := func(sql, colType string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomOnNonBigIntColumn, colType)
|
|
}
|
|
assertAddColumn := func(sql, colName string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomAlterAddColumn, colName, databaseName, tableName)
|
|
}
|
|
mustExecAndDrop := func(sql string, fns ...func()) {
|
|
tk.MustExec(sql)
|
|
for _, f := range fns {
|
|
f()
|
|
}
|
|
tk.MustExec("drop table t")
|
|
}
|
|
|
|
// Only bigint column can set auto_random
|
|
assertBigIntOnly("create table t (a char primary key auto_random(3), b int)", "char")
|
|
assertBigIntOnly("create table t (a varchar(255) primary key auto_random(3), b int)", "varchar")
|
|
assertBigIntOnly("create table t (a timestamp primary key auto_random(3), b int)", "timestamp")
|
|
|
|
// PKIsHandle, but auto_random is defined on non-primary key.
|
|
assertPKIsNotHandle("create table t (a bigint auto_random (3) primary key, b bigint auto_random (3))", "b")
|
|
assertPKIsNotHandle("create table t (a bigint auto_random (3), b bigint auto_random(3), primary key(a))", "b")
|
|
assertPKIsNotHandle("create table t (a bigint auto_random (3), b bigint auto_random(3) primary key)", "a")
|
|
|
|
// PKIsNotHandle: no primary key.
|
|
assertPKIsNotHandle("create table t (a bigint auto_random(3), b int)", "a")
|
|
// PKIsNotHandle: primary key is not a single column.
|
|
assertPKIsNotHandle("create table t (a bigint auto_random(3), b bigint, primary key (a, b))", "a")
|
|
assertPKIsNotHandle("create table t (a bigint auto_random(3), b int, c char, primary key (a, c))", "a")
|
|
|
|
// PKIsNotHandle: nonclustered integer primary key.
|
|
assertPKIsNotHandle("create table t (a bigint auto_random(3) primary key nonclustered, b int)", "a")
|
|
assertPKIsNotHandle("create table t (a bigint auto_random(3) primary key nonclustered, b int)", "a")
|
|
assertPKIsNotHandle("create table t (a int, b bigint auto_random(3) primary key nonclustered)", "b")
|
|
|
|
// Can not set auto_random along with auto_increment.
|
|
assertWithAutoInc("create table t (a bigint auto_random(3) primary key auto_increment)")
|
|
assertWithAutoInc("create table t (a bigint primary key auto_increment auto_random(3))")
|
|
assertWithAutoInc("create table t (a bigint auto_increment primary key auto_random(3))")
|
|
assertWithAutoInc("create table t (a bigint auto_random(3) auto_increment, primary key (a))")
|
|
|
|
// Can not set auto_random along with default.
|
|
assertDefault("create table t (a bigint auto_random primary key default 3)")
|
|
assertDefault("create table t (a bigint auto_random(2) primary key default 5)")
|
|
mustExecAndDrop("create table t (a bigint auto_random primary key)", func() {
|
|
assertDefault("alter table t modify column a bigint auto_random default 3")
|
|
assertDefault("alter table t alter column a set default 3")
|
|
})
|
|
|
|
// Overflow data type max length.
|
|
assertMaxOverflow("create table t (a bigint auto_random(64) primary key)", "a", 64)
|
|
assertMaxOverflow("create table t (a bigint auto_random(16) primary key)", "a", 16)
|
|
mustExecAndDrop("create table t (a bigint auto_random(5) primary key)", func() {
|
|
assertMaxOverflow("alter table t modify a bigint auto_random(64)", "a", 64)
|
|
assertMaxOverflow("alter table t modify a bigint auto_random(16)", "a", 16)
|
|
})
|
|
|
|
assertNonPositive("create table t (a bigint auto_random(0) primary key)")
|
|
tk.MustGetErrMsg("create table t (a bigint auto_random(-1) primary key)",
|
|
`[parser:1064]You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 38 near "-1) primary key)" `)
|
|
|
|
// Basic usage.
|
|
mustExecAndDrop("create table t (a bigint auto_random(1) primary key)")
|
|
mustExecAndDrop("create table t (a bigint auto_random(4) primary key)")
|
|
mustExecAndDrop("create table t (a bigint auto_random(15) primary key)")
|
|
mustExecAndDrop("create table t (a bigint primary key auto_random(4))")
|
|
mustExecAndDrop("create table t (a bigint auto_random(4), primary key (a))")
|
|
|
|
// Increase auto_random bits.
|
|
mustExecAndDrop("create table t (a bigint auto_random(5) primary key)", func() {
|
|
tk.MustExec("alter table t modify a bigint auto_random(8)")
|
|
tk.MustExec("alter table t modify a bigint auto_random(10)")
|
|
tk.MustExec("alter table t modify a bigint auto_random(12)")
|
|
})
|
|
|
|
// Auto_random can occur multiple times like other column attributes.
|
|
mustExecAndDrop("create table t (a bigint auto_random(3) auto_random(2) primary key)")
|
|
mustExecAndDrop("create table t (a bigint, b bigint auto_random(3) primary key auto_random(2))")
|
|
mustExecAndDrop("create table t (a bigint auto_random(1) auto_random(2) auto_random(3), primary key (a))")
|
|
|
|
// Add/drop the auto_random attribute is not allowed.
|
|
mustExecAndDrop("create table t (a bigint auto_random(3) primary key)", func() {
|
|
assertAlterValue("alter table t modify column a bigint")
|
|
assertAlterValue("alter table t modify column a bigint auto_random(0)")
|
|
assertAlterValue("alter table t change column a b bigint")
|
|
})
|
|
mustExecAndDrop("create table t (a bigint, b char, c bigint auto_random(3), primary key(c))", func() {
|
|
assertAlterValue("alter table t modify column c bigint")
|
|
assertAlterValue("alter table t change column c d bigint")
|
|
})
|
|
mustExecAndDrop("create table t (a bigint primary key)", func() {
|
|
assertOnlyChangeFromAutoIncPK("alter table t modify column a bigint auto_random(3)")
|
|
})
|
|
mustExecAndDrop("create table t (a bigint, b bigint, primary key(a, b))", func() {
|
|
assertOnlyChangeFromAutoIncPK("alter table t modify column a bigint auto_random(3)")
|
|
assertOnlyChangeFromAutoIncPK("alter table t modify column b bigint auto_random(3)")
|
|
})
|
|
|
|
// Add auto_random column is not allowed.
|
|
mustExecAndDrop("create table t (a bigint)", func() {
|
|
assertAddColumn("alter table t add column b int auto_random", "b")
|
|
assertAddColumn("alter table t add column b bigint auto_random", "b")
|
|
assertAddColumn("alter table t add column b bigint auto_random primary key", "b")
|
|
})
|
|
mustExecAndDrop("create table t (a bigint, b bigint primary key)", func() {
|
|
assertAddColumn("alter table t add column c int auto_random", "c")
|
|
assertAddColumn("alter table t add column c bigint auto_random", "c")
|
|
assertAddColumn("alter table t add column c bigint auto_random primary key", "c")
|
|
})
|
|
|
|
// Decrease auto_random bits is not allowed.
|
|
mustExecAndDrop("create table t (a bigint auto_random(10) primary key)", func() {
|
|
assertDecreaseBitErr("alter table t modify column a bigint auto_random(6)")
|
|
})
|
|
mustExecAndDrop("create table t (a bigint auto_random(10) primary key)", func() {
|
|
assertDecreaseBitErr("alter table t modify column a bigint auto_random(1)")
|
|
})
|
|
|
|
originStep := autoid.GetStep()
|
|
autoid.SetStep(1)
|
|
// Increase auto_random bits but it will overlap with incremental bits.
|
|
mustExecAndDrop("create table t (a bigint unsigned auto_random(5) primary key)", func() {
|
|
const alterTryCnt, rebaseOffset = 3, 1
|
|
insertSQL := fmt.Sprintf("insert into t values (%d)", ((1<<(64-10))-1)-rebaseOffset-alterTryCnt)
|
|
tk.MustExec(insertSQL)
|
|
// Try to rebase to 0..0011..1111 (54 `1`s).
|
|
tk.MustExec("alter table t modify a bigint unsigned auto_random(6)")
|
|
tk.MustExec("alter table t modify a bigint unsigned auto_random(10)")
|
|
assertOverflow("alter table t modify a bigint unsigned auto_random(11)", "a", 10, 11)
|
|
})
|
|
autoid.SetStep(originStep)
|
|
|
|
// Modifying the field type of a auto_random column is not allowed.
|
|
// Here the throw error is `ERROR 8200 (HY000): Unsupported modify column: length 11 is less than origin 20`,
|
|
// instead of `ERROR 8216 (HY000): Invalid auto random: modifying the auto_random column type is not supported`
|
|
// Because the origin column is `bigint`, it can not change to any other column type in TiDB limitation.
|
|
mustExecAndDrop("create table t (a bigint primary key auto_random(3), b int)", func() {
|
|
assertModifyColType("alter table t modify column a int auto_random(3)")
|
|
assertModifyColType("alter table t modify column a mediumint auto_random(3)")
|
|
assertModifyColType("alter table t modify column a smallint auto_random(3)")
|
|
tk.MustExec("alter table t modify column b int")
|
|
tk.MustExec("alter table t modify column b bigint")
|
|
tk.MustExec("alter table t modify column a bigint auto_random(3)")
|
|
})
|
|
|
|
// Test show warnings when create auto_random table.
|
|
assertShowWarningCorrect := func(sql string, times int) {
|
|
mustExecAndDrop(sql, func() {
|
|
note := fmt.Sprintf(autoid.AutoRandomAvailableAllocTimesNote, times)
|
|
result := fmt.Sprintf("Note|1105|%s", note)
|
|
tk.MustQuery("show warnings").Check(RowsWithSep("|", result))
|
|
c.Assert(tk.Se.GetSessionVars().StmtCtx.WarningCount(), Equals, uint16(0))
|
|
})
|
|
}
|
|
assertShowWarningCorrect("create table t (a bigint auto_random(15) primary key)", 281474976710655)
|
|
assertShowWarningCorrect("create table t (a bigint unsigned auto_random(15) primary key)", 562949953421311)
|
|
assertShowWarningCorrect("create table t (a bigint auto_random(1) primary key)", 4611686018427387903)
|
|
|
|
// Test insert into auto_random column explicitly is not allowed by default.
|
|
assertExplicitInsertDisallowed := func(sql string) {
|
|
assertInvalidAutoRandomErr(sql, autoid.AutoRandomExplicitInsertDisabledErrMsg)
|
|
}
|
|
tk.MustExec("set @@allow_auto_random_explicit_insert = false")
|
|
mustExecAndDrop("create table t (a bigint auto_random primary key)", func() {
|
|
assertExplicitInsertDisallowed("insert into t values (1)")
|
|
assertExplicitInsertDisallowed("insert into t values (3)")
|
|
tk.MustExec("insert into t values()")
|
|
})
|
|
tk.MustExec("set @@allow_auto_random_explicit_insert = true")
|
|
mustExecAndDrop("create table t (a bigint auto_random primary key)", func() {
|
|
tk.MustExec("insert into t values(1)")
|
|
tk.MustExec("insert into t values(3)")
|
|
tk.MustExec("insert into t values()")
|
|
})
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestAutoRandomChangeFromAutoInc(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("use test;")
|
|
tk.MustExec("set @@tidb_allow_remove_auto_inc = 1;")
|
|
|
|
// Basic usages.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_increment primary key);")
|
|
tk.MustExec("insert into t values (), (), ();")
|
|
tk.MustExec("alter table t modify column a bigint auto_random(3);")
|
|
tk.MustExec("insert into t values (), (), ();")
|
|
rows := tk.MustQuery("show table t next_row_id;").Rows()
|
|
c.Assert(len(rows), Equals, 1, Commentf("query result: %v", rows))
|
|
c.Assert(len(rows[0]), Equals, 5, Commentf("query result: %v", rows))
|
|
c.Assert(rows[0][4], Equals, "AUTO_RANDOM")
|
|
|
|
// Changing from auto_inc unique key is not allowed.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_increment unique key);")
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random;", errno.ErrInvalidAutoRandom)
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_increment unique key, b bigint auto_random primary key);")
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random;", errno.ErrInvalidAutoRandom)
|
|
|
|
// Changing from non-auto-inc column is not allowed.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint);")
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random;", errno.ErrInvalidAutoRandom)
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint primary key);")
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random;", errno.ErrInvalidAutoRandom)
|
|
|
|
// Changing from non BIGINT auto_inc pk column is not allowed.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a int auto_increment primary key);")
|
|
tk.MustGetErrCode("alter table t modify column a int auto_random;", errno.ErrInvalidAutoRandom)
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random;", errno.ErrInvalidAutoRandom)
|
|
|
|
// Changing from auto_random to auto_increment is not allowed.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_random primary key);")
|
|
// "Unsupported modify column: can't set auto_increment"
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_increment;", errno.ErrUnsupportedDDLOperation)
|
|
|
|
// Large auto_increment number overflows auto_random.
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_increment primary key);")
|
|
tk.MustExec("insert into t values (1<<(64-5));")
|
|
// "max allowed auto_random shard bits is 3, but got 4 on column `a`"
|
|
tk.MustGetErrCode("alter table t modify column a bigint auto_random(4);", errno.ErrInvalidAutoRandom)
|
|
tk.MustExec("drop table if exists t;")
|
|
tk.MustExec("create table t (a bigint auto_increment primary key);")
|
|
tk.MustExec("insert into t values (1<<(64-6));")
|
|
tk.MustExec("alter table t modify column a bigint auto_random(4);")
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestAutoRandomExchangePartition(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists auto_random_db")
|
|
defer tk.MustExec("drop database if exists auto_random_db")
|
|
|
|
tk.MustExec("use auto_random_db")
|
|
|
|
tk.MustExec("set @@tidb_enable_exchange_partition=1")
|
|
defer tk.MustExec("set @@tidb_enable_exchange_partition=0")
|
|
|
|
tk.MustExec("drop table if exists e1, e2, e3, e4;")
|
|
|
|
tk.MustExec("create table e1 (a bigint primary key clustered auto_random(3)) partition by hash(a) partitions 1;")
|
|
|
|
tk.MustExec("create table e2 (a bigint primary key);")
|
|
tk.MustGetErrCode("alter table e1 exchange partition p0 with table e2;", errno.ErrTablesDifferentMetadata)
|
|
|
|
tk.MustExec("create table e3 (a bigint primary key auto_random(2));")
|
|
tk.MustGetErrCode("alter table e1 exchange partition p0 with table e3;", errno.ErrTablesDifferentMetadata)
|
|
tk.MustExec("insert into e1 values (), (), ()")
|
|
|
|
tk.MustExec("create table e4 (a bigint primary key auto_random(3));")
|
|
tk.MustExec("insert into e4 values ()")
|
|
tk.MustExec("alter table e1 exchange partition p0 with table e4;")
|
|
|
|
tk.MustQuery("select count(*) from e1").Check(testkit.Rows("1"))
|
|
tk.MustExec("insert into e1 values ()")
|
|
tk.MustQuery("select count(*) from e1").Check(testkit.Rows("2"))
|
|
|
|
tk.MustQuery("select count(*) from e4").Check(testkit.Rows("3"))
|
|
tk.MustExec("insert into e4 values ()")
|
|
tk.MustQuery("select count(*) from e4").Check(testkit.Rows("4"))
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestAutoRandomIncBitsIncrementAndOffset(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists auto_random_db")
|
|
defer tk.MustExec("drop database if exists auto_random_db")
|
|
tk.MustExec("use auto_random_db")
|
|
tk.MustExec("drop table if exists t")
|
|
|
|
recreateTable := func() {
|
|
tk.MustExec("drop table if exists t")
|
|
tk.MustExec("create table t (a bigint auto_random(6) primary key clustered)")
|
|
}
|
|
truncateTable := func() {
|
|
_, _ = tk.Exec("delete from t")
|
|
}
|
|
insertTable := func() {
|
|
tk.MustExec("insert into t values ()")
|
|
}
|
|
assertIncBitsValues := func(values ...int) {
|
|
mask := strings.Repeat("1", 64-1-6)
|
|
sql := fmt.Sprintf(`select a & b'%s' from t order by a & b'%s' asc`, mask, mask)
|
|
vs := make([]string, len(values))
|
|
for i, value := range values {
|
|
vs[i] = strconv.Itoa(value)
|
|
}
|
|
tk.MustQuery(sql).Check(testkit.Rows(vs...))
|
|
}
|
|
|
|
const truncate, recreate = true, false
|
|
expect := func(vs ...int) []int { return vs }
|
|
testCase := []struct {
|
|
setupAction bool // truncate or recreate
|
|
increment int // @@auto_increment_increment
|
|
offset int // @@auto_increment_offset
|
|
results []int // the implicit allocated auto_random incremental-bit part of values
|
|
}{
|
|
{recreate, 5, 10, expect(10, 15, 20)},
|
|
{recreate, 2, 10, expect(10, 12, 14)},
|
|
{truncate, 5, 10, expect(15, 20, 25)},
|
|
{truncate, 10, 10, expect(30, 40, 50)},
|
|
{truncate, 5, 10, expect(55, 60, 65)},
|
|
}
|
|
for _, tc := range testCase {
|
|
switch tc.setupAction {
|
|
case recreate:
|
|
recreateTable()
|
|
case truncate:
|
|
truncateTable()
|
|
}
|
|
tk.Se.GetSessionVars().AutoIncrementIncrement = tc.increment
|
|
tk.Se.GetSessionVars().AutoIncrementOffset = tc.offset
|
|
for range tc.results {
|
|
insertTable()
|
|
}
|
|
assertIncBitsValues(tc.results...)
|
|
}
|
|
}
|
|
|
|
func (s *testSerialSuite) TestAutoRandomWithPreSplitRegion(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database if not exists auto_random_db;")
|
|
defer tk.MustExec("drop database if exists auto_random_db;")
|
|
tk.MustExec("use auto_random_db;")
|
|
tk.MustExec("drop table if exists t;")
|
|
|
|
origin := atomic.LoadUint32(&ddl.EnableSplitTableRegion)
|
|
atomic.StoreUint32(&ddl.EnableSplitTableRegion, 1)
|
|
defer atomic.StoreUint32(&ddl.EnableSplitTableRegion, origin)
|
|
tk.MustExec("set @@global.tidb_scatter_region=1;")
|
|
|
|
// Test pre-split table region for auto_random table.
|
|
tk.MustExec("create table t (a bigint auto_random(2) primary key clustered, b int) pre_split_regions=2;")
|
|
re := tk.MustQuery("show table t regions;")
|
|
rows := re.Rows()
|
|
c.Assert(len(rows), Equals, 4)
|
|
tbl := testGetTableByName(c, tk.Se, "auto_random_db", "t")
|
|
c.Assert(rows[1][1], Equals, fmt.Sprintf("t_%d_r_2305843009213693952", tbl.Meta().ID))
|
|
c.Assert(rows[2][1], Equals, fmt.Sprintf("t_%d_r_4611686018427387904", tbl.Meta().ID))
|
|
c.Assert(rows[3][1], Equals, fmt.Sprintf("t_%d_r_6917529027641081856", tbl.Meta().ID))
|
|
}
|
|
|
|
func (s *testSerialSuite) TestModifyingColumn4NewCollations(c *C) {
|
|
collate.SetNewCollationEnabledForTest(true)
|
|
defer collate.SetNewCollationEnabledForTest(false)
|
|
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
tk.MustExec("create database dct")
|
|
tk.MustExec("use dct")
|
|
tk.MustExec("create table t(b varchar(10) collate utf8_bin, c varchar(10) collate utf8_general_ci) collate utf8_bin")
|
|
// Column collation can be changed as long as there is no index defined.
|
|
tk.MustExec("alter table t modify b varchar(10) collate utf8_general_ci")
|
|
tk.MustExec("alter table t modify c varchar(10) collate utf8_bin")
|
|
tk.MustExec("alter table t modify c varchar(10) collate utf8_unicode_ci")
|
|
tk.MustExec("alter table t charset utf8 collate utf8_general_ci")
|
|
tk.MustExec("alter table t convert to charset utf8 collate utf8_bin")
|
|
tk.MustExec("alter table t convert to charset utf8 collate utf8_unicode_ci")
|
|
tk.MustExec("alter table t convert to charset utf8 collate utf8_general_ci")
|
|
tk.MustExec("alter table t modify b varchar(10) collate utf8_unicode_ci")
|
|
tk.MustExec("alter table t modify b varchar(10) collate utf8_bin")
|
|
|
|
tk.MustExec("alter table t add index b_idx(b)")
|
|
tk.MustExec("alter table t add index c_idx(c)")
|
|
tk.MustGetErrMsg("alter table t modify b varchar(10) collate utf8_general_ci", "[ddl:8200]Unsupported modifying collation of column 'b' from 'utf8_bin' to 'utf8_general_ci' when index is defined on it.")
|
|
tk.MustGetErrMsg("alter table t modify c varchar(10) collate utf8_bin", "[ddl:8200]Unsupported modifying collation of column 'c' from 'utf8_general_ci' to 'utf8_bin' when index is defined on it.")
|
|
tk.MustGetErrMsg("alter table t modify c varchar(10) collate utf8_unicode_ci", "[ddl:8200]Unsupported modifying collation of column 'c' from 'utf8_general_ci' to 'utf8_unicode_ci' when index is defined on it.")
|
|
tk.MustGetErrMsg("alter table t convert to charset utf8 collate utf8_general_ci", "[ddl:8200]Unsupported converting collation of column 'b' from 'utf8_bin' to 'utf8_general_ci' when index is defined on it.")
|
|
// Change to a compatible collation is allowed.
|
|
tk.MustExec("alter table t modify c varchar(10) collate utf8mb4_general_ci")
|
|
// Change the default collation of table is allowed.
|
|
tk.MustExec("alter table t collate utf8mb4_general_ci")
|
|
tk.MustExec("alter table t charset utf8mb4 collate utf8mb4_bin")
|
|
tk.MustExec("alter table t charset utf8mb4 collate utf8mb4_unicode_ci")
|
|
tk.MustExec("alter table t charset utf8mb4 collate utf8mb4_zh_pinyin_tidb_as_cs")
|
|
// Change the default collation of database is allowed.
|
|
tk.MustExec("alter database dct charset utf8mb4 collate utf8mb4_general_ci")
|
|
}
|
|
|
|
func (s *testSerialSuite) TestForbidUnsupportedCollations(c *C) {
|
|
collate.SetNewCollationEnabledForTest(true)
|
|
defer collate.SetNewCollationEnabledForTest(false)
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
|
|
mustGetUnsupportedCollation := func(sql string, coll string) {
|
|
tk.MustGetErrMsg(sql, fmt.Sprintf("[ddl:1273]Unsupported collation when new collation is enabled: '%s'", coll))
|
|
}
|
|
// Test default collation of database.
|
|
mustGetUnsupportedCollation("create database ucd charset utf8mb4 collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("create database ucd charset utf8 collate utf8_roman_ci", "utf8_roman_ci")
|
|
tk.MustExec("create database ucd")
|
|
mustGetUnsupportedCollation("alter database ucd charset utf8mb4 collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("alter database ucd collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
|
|
// Test default collation of table.
|
|
tk.MustExec("use ucd")
|
|
mustGetUnsupportedCollation("create table t(a varchar(20)) charset utf8mb4 collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("create table t(a varchar(20)) collate utf8_roman_ci", "utf8_roman_ci")
|
|
tk.MustExec("create table t(a varchar(20)) collate utf8mb4_general_ci")
|
|
mustGetUnsupportedCollation("alter table t default collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("alter table t convert to charset utf8mb4 collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
|
|
// Test collation of columns.
|
|
mustGetUnsupportedCollation("create table t1(a varchar(20)) collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("create table t1(a varchar(20)) charset utf8 collate utf8_roman_ci", "utf8_roman_ci")
|
|
tk.MustExec("create table t1(a varchar(20))")
|
|
mustGetUnsupportedCollation("alter table t1 modify a varchar(20) collate utf8mb4_roman_ci", "utf8mb4_roman_ci")
|
|
mustGetUnsupportedCollation("alter table t1 modify a varchar(20) charset utf8 collate utf8_roman_ci", "utf8_roman_ci")
|
|
mustGetUnsupportedCollation("alter table t1 modify a varchar(20) charset utf8 collate utf8_roman_ci", "utf8_roman_ci")
|
|
|
|
// TODO(bb7133): fix the following cases by setting charset from collate firstly.
|
|
// mustGetUnsupportedCollation("create database ucd collate utf8mb4_unicode_ci", errMsgUnsupportedUnicodeCI)
|
|
// mustGetUnsupportedCollation("alter table t convert to collate utf8mb4_unicode_ci", "utf8mb4_unicode_ci")
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestInvisibleIndex(c *C) {
|
|
tk := testkit.NewTestKit(c, s.store)
|
|
|
|
tk.MustExec("use test")
|
|
tk.MustExec("drop table if exists t,t1,t2,t3,t4,t5,t6")
|
|
|
|
// The DDL statement related to invisible index.
|
|
showIndexes := "select index_name, is_visible from information_schema.statistics where table_schema = 'test' and table_name = 't'"
|
|
// 1. Create table with invisible index
|
|
tk.MustExec("create table t (a int, b int, unique (a) invisible)")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows("a NO"))
|
|
tk.MustExec("insert into t values (1, 2)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2"))
|
|
// 2. Drop invisible index
|
|
tk.MustExec("alter table t drop index a")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows())
|
|
tk.MustExec("insert into t values (3, 4)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2", "3 4"))
|
|
// 3. Add an invisible index
|
|
tk.MustExec("alter table t add index (b) invisible")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows("b NO"))
|
|
tk.MustExec("insert into t values (5, 6)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2", "3 4", "5 6"))
|
|
// 4. Drop it
|
|
tk.MustExec("alter table t drop index b")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows())
|
|
tk.MustExec("insert into t values (7, 8)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2", "3 4", "5 6", "7 8"))
|
|
// 5. Create a multiple-column invisible index
|
|
tk.MustExec("alter table t add index a_b(a, b) invisible")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows("a_b NO", "a_b NO"))
|
|
tk.MustExec("insert into t values (9, 10)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2", "3 4", "5 6", "7 8", "9 10"))
|
|
// 6. Drop it
|
|
tk.MustExec("alter table t drop index a_b")
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows())
|
|
tk.MustExec("insert into t values (11, 12)")
|
|
tk.MustQuery("select * from t").Check(testkit.Rows("1 2", "3 4", "5 6", "7 8", "9 10", "11 12"))
|
|
|
|
// Limitation: Primary key cannot be invisible index
|
|
tk.MustGetErrCode("create table t1 (a int, primary key (a) nonclustered invisible)", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustGetErrCode("create table t1 (a int, b int, primary key (a, b) nonclustered invisible)", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustExec("create table t1 (a int, b int)")
|
|
tk.MustGetErrCode("alter table t1 add primary key(a) nonclustered invisible", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustGetErrCode("alter table t1 add primary key(a, b) nonclustered invisible", errno.ErrPKIndexCantBeInvisible)
|
|
|
|
// Implicit primary key cannot be invisible index
|
|
// Create a implicit primary key
|
|
tk.MustGetErrCode("create table t2(a int not null, unique (a) invisible)", errno.ErrPKIndexCantBeInvisible)
|
|
// Column `a` become implicit primary key after DDL statement on itself
|
|
tk.MustExec("create table t2(a int not null)")
|
|
tk.MustGetErrCode("alter table t2 add unique (a) invisible", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustExec("create table t3(a int, unique index (a) invisible)")
|
|
tk.MustGetErrCode("alter table t3 modify column a int not null", errno.ErrPKIndexCantBeInvisible)
|
|
// Only first unique column can be implicit primary
|
|
tk.MustExec("create table t4(a int not null, b int not null, unique (a), unique (b) invisible)")
|
|
showIndexes = "select index_name, is_visible from information_schema.statistics where table_schema = 'test' and table_name = 't4'"
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows("a YES", "b NO"))
|
|
tk.MustExec("insert into t4 values (1, 2)")
|
|
tk.MustQuery("select * from t4").Check(testkit.Rows("1 2"))
|
|
tk.MustGetErrCode("create table t5(a int not null, b int not null, unique (b) invisible, unique (a))", errno.ErrPKIndexCantBeInvisible)
|
|
// Column `b` become implicit primary key after DDL statement on other columns
|
|
tk.MustExec("create table t5(a int not null, b int not null, unique (a), unique (b) invisible)")
|
|
tk.MustGetErrCode("alter table t5 drop index a", errno.ErrPKIndexCantBeInvisible)
|
|
tk.MustGetErrCode("alter table t5 modify column a int null", errno.ErrPKIndexCantBeInvisible)
|
|
// If these is a explicit primary key, no key will become implicit primary key
|
|
tk.MustExec("create table t6 (a int not null, b int, unique (a) invisible, primary key(b) nonclustered)")
|
|
showIndexes = "select index_name, is_visible from information_schema.statistics where table_schema = 'test' and table_name = 't6'"
|
|
tk.MustQuery(showIndexes).Check(testkit.Rows("a NO", "PRIMARY YES"))
|
|
tk.MustExec("insert into t6 values (1, 2)")
|
|
tk.MustQuery("select * from t6").Check(testkit.Rows("1 2"))
|
|
tk.MustGetErrCode("alter table t6 drop primary key", errno.ErrPKIndexCantBeInvisible)
|
|
res := tk.MustQuery("show index from t6 where Key_name='PRIMARY';")
|
|
c.Check(len(res.Rows()), Equals, 1)
|
|
}
|
|
|
|
func (s *testIntegrationSuite7) TestCreateClusteredIndex(c *C) {
|
|
tk := testkit.NewTestKitWithInit(c, s.store)
|
|
tk.Se.GetSessionVars().EnableClusteredIndex = variable.ClusteredIndexDefModeOn
|
|
tk.MustExec("CREATE TABLE t1 (a int primary key, b int)")
|
|
tk.MustExec("CREATE TABLE t2 (a varchar(255) primary key, b int)")
|
|
tk.MustExec("CREATE TABLE t3 (a int, b int, c int, primary key (a, b))")
|
|
tk.MustExec("CREATE TABLE t4 (a int, b int, c int)")
|
|
ctx := tk.Se.(sessionctx.Context)
|
|
is := domain.GetDomain(ctx).InfoSchema()
|
|
tbl, err := is.TableByName(model.NewCIStr("test"), model.NewCIStr("t1"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().PKIsHandle, IsTrue)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsFalse)
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t2"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsTrue)
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t3"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsTrue)
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t4"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsFalse)
|
|
|
|
tk.MustExec("CREATE TABLE t5 (a varchar(255) primary key nonclustered, b int)")
|
|
tk.MustExec("CREATE TABLE t6 (a int, b int, c int, primary key (a, b) nonclustered)")
|
|
is = domain.GetDomain(ctx).InfoSchema()
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t5"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsFalse)
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t6"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsFalse)
|
|
|
|
tk.MustExec("CREATE TABLE t21 like t2")
|
|
tk.MustExec("CREATE TABLE t31 like t3")
|
|
is = domain.GetDomain(ctx).InfoSchema()
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t21"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsTrue)
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t31"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsTrue)
|
|
|
|
tk.Se.GetSessionVars().EnableClusteredIndex = variable.ClusteredIndexDefModeIntOnly
|
|
tk.MustExec("CREATE TABLE t7 (a varchar(255) primary key, b int)")
|
|
is = domain.GetDomain(ctx).InfoSchema()
|
|
tbl, err = is.TableByName(model.NewCIStr("test"), model.NewCIStr("t7"))
|
|
c.Assert(err, IsNil)
|
|
c.Assert(tbl.Meta().IsCommonHandle, IsFalse)
|
|
}
|
|
|
|
func (s *testSerialDBSuite) TestCreateTableNoBlock(c *C) {
|
|
tk := testkit.NewTestKitWithInit(c, s.store)
|
|
c.Assert(failpoint.Enable("github.com/pingcap/tidb/ddl/checkOwnerCheckAllVersionsWaitTime", `return(true)`), IsNil)
|
|
defer func() {
|
|
c.Assert(failpoint.Disable("github.com/pingcap/tidb/ddl/checkOwnerCheckAllVersionsWaitTime"), IsNil)
|
|
}()
|
|
save := variable.GetDDLErrorCountLimit()
|
|
variable.SetDDLErrorCountLimit(1)
|
|
defer func() {
|
|
variable.SetDDLErrorCountLimit(save)
|
|
}()
|
|
|
|
tk.MustExec("drop table if exists t")
|
|
_, err := tk.Exec("create table t(a int)")
|
|
c.Assert(err, NotNil)
|
|
}
|
|
|
|
func (s *testSerialSuite) TestCheckEnumLength(c *C) {
|
|
tk := testkit.NewTestKitWithInit(c, s.store)
|
|
tk.MustExec("drop table if exists t1,t2,t3,t4,t5")
|
|
tk.MustGetErrCode("create table t1 (a enum('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))", errno.ErrTooLongValueForType)
|
|
tk.MustGetErrCode("create table t1 (a set('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))", errno.ErrTooLongValueForType)
|
|
tk.MustExec("create table t2 (id int primary key)")
|
|
tk.MustGetErrCode("alter table t2 add a enum('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa')", errno.ErrTooLongValueForType)
|
|
tk.MustGetErrCode("alter table t2 add a set('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa')", errno.ErrTooLongValueForType)
|
|
config.UpdateGlobal(func(conf *config.Config) {
|
|
conf.EnableEnumLengthLimit = false
|
|
})
|
|
_, err := tk.Exec("create table t3 (a enum('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))")
|
|
c.Assert(err, IsNil)
|
|
tk.MustExec("insert into t3 values(1)")
|
|
tk.MustQuery("select a from t3").Check(testkit.Rows("aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"))
|
|
_, err = tk.Exec("create table t4 (a set('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))")
|
|
c.Assert(err, IsNil)
|
|
|
|
config.UpdateGlobal(func(conf *config.Config) {
|
|
conf.EnableEnumLengthLimit = true
|
|
})
|
|
tk.MustGetErrCode("create table t5 (a enum('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))", errno.ErrTooLongValueForType)
|
|
tk.MustGetErrCode("create table t5 (a set('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'))", errno.ErrTooLongValueForType)
|
|
tk.MustExec("drop table if exists t1,t2,t3,t4,t5")
|
|
}
|