Files
tidb/pkg/planner/core/planbuilder.go

6518 lines
252 KiB
Go

// Copyright 2015 PingCAP, Inc.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package core
import (
"context"
"encoding/binary"
"fmt"
"maps"
"math"
"net/url"
"reflect"
"regexp"
"slices"
"strconv"
"strings"
"github.com/docker/go-units"
"github.com/pingcap/errors"
"github.com/pingcap/tidb/pkg/bindinfo"
"github.com/pingcap/tidb/pkg/config"
"github.com/pingcap/tidb/pkg/config/kerneltype"
"github.com/pingcap/tidb/pkg/domain"
"github.com/pingcap/tidb/pkg/expression"
"github.com/pingcap/tidb/pkg/infoschema"
"github.com/pingcap/tidb/pkg/kv"
"github.com/pingcap/tidb/pkg/meta/metadef"
"github.com/pingcap/tidb/pkg/meta/model"
"github.com/pingcap/tidb/pkg/objstore"
"github.com/pingcap/tidb/pkg/objstore/s3like"
"github.com/pingcap/tidb/pkg/parser"
"github.com/pingcap/tidb/pkg/parser/ast"
"github.com/pingcap/tidb/pkg/parser/auth"
"github.com/pingcap/tidb/pkg/parser/charset"
"github.com/pingcap/tidb/pkg/parser/mysql"
"github.com/pingcap/tidb/pkg/parser/opcode"
"github.com/pingcap/tidb/pkg/parser/terror"
"github.com/pingcap/tidb/pkg/planner/core/base"
"github.com/pingcap/tidb/pkg/planner/core/operator/logicalop"
"github.com/pingcap/tidb/pkg/planner/core/operator/physicalop"
"github.com/pingcap/tidb/pkg/planner/core/resolve"
"github.com/pingcap/tidb/pkg/planner/core/rule"
"github.com/pingcap/tidb/pkg/planner/property"
"github.com/pingcap/tidb/pkg/planner/util"
"github.com/pingcap/tidb/pkg/planner/util/domainmisc"
"github.com/pingcap/tidb/pkg/privilege"
"github.com/pingcap/tidb/pkg/session/syssession"
"github.com/pingcap/tidb/pkg/sessionctx"
"github.com/pingcap/tidb/pkg/sessionctx/stmtctx"
"github.com/pingcap/tidb/pkg/sessionctx/vardef"
"github.com/pingcap/tidb/pkg/sessionctx/variable"
"github.com/pingcap/tidb/pkg/sessiontxn/staleread"
"github.com/pingcap/tidb/pkg/statistics"
statslogutil "github.com/pingcap/tidb/pkg/statistics/handle/logutil"
handleutil "github.com/pingcap/tidb/pkg/statistics/handle/util"
"github.com/pingcap/tidb/pkg/table"
"github.com/pingcap/tidb/pkg/table/tables"
"github.com/pingcap/tidb/pkg/types"
driver "github.com/pingcap/tidb/pkg/types/parser_driver"
"github.com/pingcap/tidb/pkg/util/chunk"
"github.com/pingcap/tidb/pkg/util/dbterror"
"github.com/pingcap/tidb/pkg/util/dbterror/exeerrors"
"github.com/pingcap/tidb/pkg/util/dbterror/plannererrors"
"github.com/pingcap/tidb/pkg/util/execdetails"
"github.com/pingcap/tidb/pkg/util/hint"
"github.com/pingcap/tidb/pkg/util/intest"
"github.com/pingcap/tidb/pkg/util/logutil"
utilparser "github.com/pingcap/tidb/pkg/util/parser"
"github.com/pingcap/tidb/pkg/util/ranger"
semv1 "github.com/pingcap/tidb/pkg/util/sem"
sem "github.com/pingcap/tidb/pkg/util/sem/compat"
semv2 "github.com/pingcap/tidb/pkg/util/sem/v2"
"github.com/pingcap/tidb/pkg/util/set"
"github.com/pingcap/tidb/pkg/util/sqlexec"
"github.com/tikv/client-go/v2/oracle"
"go.uber.org/zap"
)
type visitInfo struct {
privilege mysql.PrivilegeType
db string
table string
column string
err error
alterWritable bool
// if multiple privileges is provided, user should
// have at least one privilege to pass the check.
dynamicPrivs []string
dynamicWithGrant bool
}
func (v *visitInfo) Equals(other *visitInfo) bool {
return v.privilege == other.privilege &&
v.db == other.db &&
v.table == other.table &&
v.column == other.column &&
v.err == other.err &&
v.alterWritable == other.alterWritable &&
reflect.DeepEqual(v.dynamicPrivs, other.dynamicPrivs) &&
v.dynamicWithGrant == other.dynamicWithGrant
}
// clauseCode indicates in which clause the column is currently.
type clauseCode int
const (
unknowClause clauseCode = iota
fieldList
havingClause
onClause
orderByClause
whereClause
groupByClause
showStatement
globalOrderByClause
expressionClause
windowOrderByClause
partitionByClause
)
var clauseMsg = map[clauseCode]string{
unknowClause: "",
fieldList: "field list",
havingClause: "having clause",
onClause: "on clause",
orderByClause: "order clause",
whereClause: "where clause",
groupByClause: "group statement",
showStatement: "show statement",
globalOrderByClause: "global ORDER clause",
expressionClause: "expression",
windowOrderByClause: "window order by",
partitionByClause: "window partition by",
}
type capFlagType = uint64
const (
_ capFlagType = iota
// canExpandAST indicates whether the origin AST can be expanded during plan
// building. ONLY used for `CreateViewStmt` now.
canExpandAST
// renameView indicates a view is being renamed, so we cannot use the origin
// definition of that view.
renameView
)
type cteInfo struct {
def *ast.CommonTableExpression
// nonRecursive is used to decide if a CTE is visible. If a CTE start with `WITH RECURSIVE`, then nonRecursive is false,
// so it is visible in its definition.
nonRecursive bool
// useRecursive is used to record if a subSelect in CTE's definition refer to itself. This help us to identify the seed part and recursive part.
useRecursive bool
isBuilding bool
// isDistinct indicates if the union between seed part and recursive part is distinct or not.
isDistinct bool
// seedLP is the seed part's logical plan.
seedLP base.LogicalPlan
// recurLP is the recursive part's logical plan.
recurLP base.LogicalPlan
// storageID for this CTE.
storageID int
// optFlag is the optFlag for the whole CTE.
optFlag uint64
// enterSubquery and recursiveRef are used to check "recursive table must be referenced only once, and not in any subquery".
enterSubquery bool
recursiveRef bool
limitLP base.LogicalPlan
// seedStat is shared between logicalCTE and logicalCTETable.
seedStat *property.StatsInfo
// The LogicalCTEs that reference the same table should share the same CteClass.
cteClass *logicalop.CTEClass
// isInline will determine whether it can be inlined when **CTE is used**
isInline bool
// forceInlineByHintOrVar will be true when CTE is hint by merge() or session variable "tidb_opt_force_inline_cte=true"
forceInlineByHintOrVar bool
// If CTE contain aggregation, window function, order by, distinct and limit in query (Indirect references to other cte containing those operator in the query are also counted.)
containRecursiveForbiddenOperator bool
// Compute in preprocess phase. Record how many consumers the current CTE has
consumerCount int
}
type subQueryCtx = uint64
const (
notHandlingSubquery subQueryCtx = iota
handlingExistsSubquery
handlingCompareSubquery
handlingInSubquery
handlingScalarSubquery
)
// PlanBuilder builds Plan from an ast.Node.
// It just builds the ast node straightforwardly.
type PlanBuilder struct {
ctx base.PlanContext
is infoschema.InfoSchema
outerSchemas []*expression.Schema
outerNames [][]*types.FieldName
outerCTEs []*cteInfo
// outerBlockExpand register current Expand OP for rollup syntax in every select query block.
outerBlockExpand []*logicalop.LogicalExpand
currentBlockExpand *logicalop.LogicalExpand
// colMapper stores the column that must be pre-resolved.
colMapper map[*ast.ColumnNameExpr]int
// visitInfo is used for privilege check.
visitInfo []visitInfo
tableHintInfo []*hint.PlanHints
// optFlag indicates the flags of the optimizer rules.
optFlag uint64
// capFlag indicates the capability flags.
capFlag capFlagType
curClause clauseCode
// rewriterPool stores the expressionRewriter we have created to reuse it if it has been released.
// rewriterCounter counts how many rewriter is being used.
rewriterPool []*expressionRewriter
rewriterCounter int
windowSpecs map[string]*ast.WindowSpec
inUpdateStmt bool
inDeleteStmt bool
// inStraightJoin represents whether the current "SELECT" statement has
// "STRAIGHT_JOIN" option.
inStraightJoin bool
// handleHelper records the handle column position for tables. Delete/Update/SelectLock/UnionScan may need this information.
// It collects the information by the following procedure:
// Since we build the plan tree from bottom to top, we maintain a stack to record the current handle information.
// If it's a dataSource/tableDual node, we create a new map.
// If it's a aggregation, we pop the map and push a nil map since no handle information left.
// If it's a union, we pop all children's and push a nil map.
// If it's a join, we pop its children's out then merge them and push the new map to stack.
// If we meet a subquery or CTE, it's clearly that it's an independent problem so we just pop one map out when we
// finish building the subquery or CTE.
handleHelper *handleColHelper
hintProcessor *hint.QBHintHandler
// qbOffset is the offsets of current processing select stmts.
qbOffset []int
// SelectLock need this information to locate the lock on partitions.
partitionedTable []table.PartitionedTable
// buildingViewStack is used to check whether there is a recursive view.
buildingViewStack set.StringSet
// renamingViewName is the name of the view which is being renamed.
renamingViewName string
// isCreateView indicates whether the query is create view.
isCreateView bool
// evalDefaultExpr needs this information to find the corresponding column.
// It stores the OutputNames before buildProjection.
allNames [][]*types.FieldName
// isSampling indicates whether the query is sampling.
isSampling bool
// correlatedAggMapper stores columns for correlated aggregates which should be evaluated in outer query.
correlatedAggMapper map[*ast.AggregateFuncExpr]*expression.CorrelatedColumn
// isForUpdateRead should be true in either of the following situations
// 1. use `inside insert`, `update`, `delete` or `select for update` statement
// 2. isolation level is RC
isForUpdateRead bool
allocIDForCTEStorage int
buildingRecursivePartForCTE bool
buildingCTE bool
// Check whether the current building query is a CTE
isCTE bool
// CTE table name in lower case, it can be nil
nameMapCTE map[string]struct{}
// subQueryCtx and subQueryHintFlags are for handling subquery related hints.
// Note: "subquery" here only contains subqueries that are handled by the expression rewriter, i.e., [NOT] IN,
// [NOT] EXISTS, compare + ANY/ALL and scalar subquery. Derived table doesn't belong to this.
// We need these fields to passing information because:
// 1. We know what kind of subquery is this only when we're in the outer query block.
// 2. We know if there are such hints only when we're in the subquery block.
// 3. These hints are only applicable when they are in a subquery block. And for some hints, they are only
// applicable for some kinds of subquery.
// 4. If a hint is set and is applicable, the corresponding logic is handled in the outer query block.
// Example SQL: select * from t where exists(select /*+ SEMI_JOIN_REWRITE() */ 1 from t1 where t.a=t1.a)
// subQueryCtx indicates if we are handling a subquery, and what kind of subquery is it.
subQueryCtx subQueryCtx
// subQueryHintFlags stores subquery related hints that are set and applicable in the query block.
// It's for returning information to buildSubquery().
subQueryHintFlags uint64 // TODO: move this field to hint.PlanHints
// disableSubQueryPreprocessing indicates whether to pre-process uncorrelated sub-queries in rewriting stage.
disableSubQueryPreprocessing bool
// noDecorrelate indicates whether decorrelation should be disabled for correlated aggregates in subqueries
noDecorrelate bool
// allowBuildCastArray indicates whether allow cast(... as ... array).
allowBuildCastArray bool
// resolveCtx is set when calling Build, it's only effective in the current Build call.
resolveCtx *resolve.Context
}
type handleColHelper struct {
id2HandleMapStack []map[int64][]util.HandleCols
stackTail int
}
func (hch *handleColHelper) resetForReuse() {
*hch = handleColHelper{
id2HandleMapStack: hch.id2HandleMapStack[:0],
}
}
func (hch *handleColHelper) popMap() map[int64][]util.HandleCols {
ret := hch.id2HandleMapStack[hch.stackTail-1]
hch.stackTail--
hch.id2HandleMapStack = hch.id2HandleMapStack[:hch.stackTail]
return ret
}
func (hch *handleColHelper) pushMap(m map[int64][]util.HandleCols) {
hch.id2HandleMapStack = append(hch.id2HandleMapStack, m)
hch.stackTail++
}
func (hch *handleColHelper) mergeAndPush(m1, m2 map[int64][]util.HandleCols) {
newMap := make(map[int64][]util.HandleCols, max(len(m1), len(m2)))
for k, v := range m1 {
newMap[k] = make([]util.HandleCols, len(v))
copy(newMap[k], v)
}
for k, v := range m2 {
if _, ok := newMap[k]; ok {
newMap[k] = append(newMap[k], v...)
} else {
newMap[k] = make([]util.HandleCols, len(v))
copy(newMap[k], v)
}
}
hch.pushMap(newMap)
}
func (hch *handleColHelper) tailMap() map[int64][]util.HandleCols {
return hch.id2HandleMapStack[hch.stackTail-1]
}
// GetVisitInfo gets the visitInfo of the PlanBuilder.
func (b *PlanBuilder) GetVisitInfo() []visitInfo {
return b.visitInfo
}
// GetIsForUpdateRead gets if the PlanBuilder use forUpdateRead
func (b *PlanBuilder) GetIsForUpdateRead() bool {
return b.isForUpdateRead
}
// GetDBTableInfo gets the accessed dbs and tables info.
func GetDBTableInfo(visitInfo []visitInfo) []stmtctx.TableEntry {
var tables []stmtctx.TableEntry
existsFunc := func(tbls []stmtctx.TableEntry, tbl *stmtctx.TableEntry) bool {
return slices.Contains(tbls, *tbl)
}
for _, v := range visitInfo {
if v.db == "" && v.table == "" {
// when v.db == "" and v.table == "", it means this visitInfo is for dynamic privilege,
// so it is not related to any database or table.
continue
}
tbl := &stmtctx.TableEntry{DB: v.db, Table: v.table}
if !existsFunc(tables, tbl) {
tables = append(tables, *tbl)
}
}
return tables
}
// GetOptFlag gets the OptFlag of the PlanBuilder.
func (b *PlanBuilder) GetOptFlag() uint64 {
if b.isSampling {
// Disable logical optimization to avoid the optimizer
// push down/eliminate operands like Selection, Limit or Sort.
return 0
}
return b.optFlag
}
func (b *PlanBuilder) getSelectOffset() int {
if len(b.qbOffset) > 0 {
return b.qbOffset[len(b.qbOffset)-1]
}
return -1
}
func (b *PlanBuilder) pushSelectOffset(offset int) {
b.qbOffset = append(b.qbOffset, offset)
}
func (b *PlanBuilder) popSelectOffset() {
b.qbOffset = b.qbOffset[:len(b.qbOffset)-1]
}
// PlanBuilderOpt is used to adjust the plan builder.
type PlanBuilderOpt interface {
Apply(builder *PlanBuilder)
}
// PlanBuilderOptNoExecution means the plan builder should not run any executor during Build().
type PlanBuilderOptNoExecution struct{}
// Apply implements the interface PlanBuilderOpt.
func (PlanBuilderOptNoExecution) Apply(builder *PlanBuilder) {
builder.disableSubQueryPreprocessing = true
}
// PlanBuilderOptAllowCastArray means the plan builder should allow build cast(... as ... array).
type PlanBuilderOptAllowCastArray struct{}
// Apply implements the interface PlanBuilderOpt.
func (PlanBuilderOptAllowCastArray) Apply(builder *PlanBuilder) {
builder.allowBuildCastArray = true
}
// NewPlanBuilder creates a new PlanBuilder.
func NewPlanBuilder(opts ...PlanBuilderOpt) *PlanBuilder {
builder := &PlanBuilder{
outerCTEs: make([]*cteInfo, 0),
colMapper: make(map[*ast.ColumnNameExpr]int),
handleHelper: &handleColHelper{id2HandleMapStack: make([]map[int64][]util.HandleCols, 0)},
correlatedAggMapper: make(map[*ast.AggregateFuncExpr]*expression.CorrelatedColumn),
}
for _, opt := range opts {
opt.Apply(builder)
}
return builder
}
// Init initialize a PlanBuilder.
// Return the original PlannerSelectBlockAsName as well, callers decide if
// PlannerSelectBlockAsName should be restored after using this builder.
// This is The comman code pattern to use it:
// NewPlanBuilder().Init(sctx, is, processor)
func (b *PlanBuilder) Init(sctx base.PlanContext, is infoschema.InfoSchema, processor *hint.QBHintHandler) (*PlanBuilder, []ast.HintTable) {
savedBlockNames := sctx.GetSessionVars().PlannerSelectBlockAsName.Load()
if processor == nil {
sctx.GetSessionVars().PlannerSelectBlockAsName.Store(&[]ast.HintTable{})
} else {
newPlannerSelectBlockAsName := make([]ast.HintTable, processor.MaxSelectStmtOffset()+1)
sctx.GetSessionVars().PlannerSelectBlockAsName.Store(&newPlannerSelectBlockAsName)
}
b.ctx = sctx
b.is = is
b.hintProcessor = processor
b.isForUpdateRead = sctx.GetSessionVars().IsPessimisticReadConsistency()
b.noDecorrelate = sctx.GetSessionVars().EnableNoDecorrelateInSelect
if savedBlockNames == nil {
return b, nil
}
return b, *savedBlockNames
}
// ResetForReuse reset the plan builder, put it into pool for reuse.
// After reset for reuse, the object should be equal to a object returned by NewPlanBuilder().
func (b *PlanBuilder) ResetForReuse() *PlanBuilder {
// Save some fields for reuse.
saveOuterCTEs := b.outerCTEs[:0]
saveColMapper := b.colMapper
for k := range saveColMapper {
delete(saveColMapper, k)
}
saveHandleHelper := b.handleHelper
saveHandleHelper.resetForReuse()
saveCorrelateAggMapper := b.correlatedAggMapper
for k := range saveCorrelateAggMapper {
delete(saveCorrelateAggMapper, k)
}
// Reset ALL the fields.
*b = PlanBuilder{}
// Reuse part of the fields.
// It's a bit conservative but easier to get right.
b.outerCTEs = saveOuterCTEs
b.colMapper = saveColMapper
b.handleHelper = saveHandleHelper
b.correlatedAggMapper = saveCorrelateAggMapper
b.noDecorrelate = false
// Add more fields if they are safe to be reused.
return b
}
// Build builds the ast node to a Plan.
func (b *PlanBuilder) Build(ctx context.Context, node *resolve.NodeW) (base.Plan, error) {
err := b.checkSEMStmt(node.Node)
if err != nil {
return nil, err
}
// Build might be called recursively, right now they all share the same resolve
// context, so it's ok to override it.
b.resolveCtx = node.GetResolveContext()
b.optFlag |= rule.FlagPruneColumns
switch x := node.Node.(type) {
case *ast.AdminStmt:
return b.buildAdmin(ctx, x)
case *ast.DeallocateStmt:
return &Deallocate{Name: x.Name}, nil
case *ast.DeleteStmt:
return b.buildDelete(ctx, x)
case *ast.ExecuteStmt:
return b.buildExecute(ctx, x)
case *ast.ExplainStmt:
return b.buildExplain(ctx, x)
case *ast.ExplainForStmt:
return b.buildExplainFor(x)
case *ast.TraceStmt:
return b.buildTrace(x)
case *ast.InsertStmt:
return b.buildInsert(ctx, x)
case *ast.ImportIntoStmt:
return b.buildImportInto(ctx, x)
case *ast.LoadDataStmt:
return b.buildLoadData(ctx, x)
case *ast.LoadStatsStmt:
return b.buildLoadStats(x), nil
case *ast.RefreshStatsStmt:
return b.buildRefreshStats(x)
case *ast.LockStatsStmt:
return b.buildLockStats(x), nil
case *ast.UnlockStatsStmt:
return b.buildUnlockStats(x), nil
case *ast.PlanReplayerStmt:
return b.buildPlanReplayer(x), nil
case *ast.TrafficStmt:
return b.buildTraffic(x), nil
case *ast.PrepareStmt:
return b.buildPrepare(x), nil
case *ast.SelectStmt:
if x.SelectIntoOpt != nil {
return b.buildSelectInto(ctx, x)
}
return b.buildSelect(ctx, x)
case *ast.SetOprStmt:
return b.buildSetOpr(ctx, x)
case *ast.UpdateStmt:
return b.buildUpdate(ctx, x)
case *ast.ShowStmt:
return b.buildShow(ctx, x)
case *ast.DoStmt:
return b.buildDo(ctx, x)
case *ast.SetStmt:
return b.buildSet(ctx, x)
case *ast.SetConfigStmt:
return b.buildSetConfig(ctx, x)
case *ast.AnalyzeTableStmt:
return b.buildAnalyze(x)
case *ast.BinlogStmt, *ast.FlushStmt, *ast.UseStmt, *ast.BRIEStmt,
*ast.BeginStmt, *ast.CommitStmt, *ast.SavepointStmt, *ast.ReleaseSavepointStmt, *ast.RollbackStmt, *ast.CreateUserStmt, *ast.SetPwdStmt, *ast.AlterInstanceStmt,
*ast.GrantStmt, *ast.DropUserStmt, *ast.AlterUserStmt, *ast.AlterRangeStmt, *ast.RevokeStmt, *ast.KillStmt, *ast.DropStatsStmt,
*ast.GrantRoleStmt, *ast.RevokeRoleStmt, *ast.SetRoleStmt, *ast.SetDefaultRoleStmt, *ast.ShutdownStmt,
*ast.RenameUserStmt, *ast.NonTransactionalDMLStmt, *ast.SetSessionStatesStmt, *ast.SetResourceGroupStmt, *ast.CancelDistributionJobStmt,
*ast.ImportIntoActionStmt, *ast.CalibrateResourceStmt, *ast.AddQueryWatchStmt, *ast.DropQueryWatchStmt, *ast.DropProcedureStmt:
return b.buildSimple(ctx, node.Node.(ast.StmtNode))
case ast.DDLNode:
if b.ctx.IsCrossKS() {
return nil, errors.New("DDL is not supported in cross keyspace session")
}
return b.buildDDL(ctx, x)
case *ast.CreateBindingStmt:
return b.buildCreateBindPlan(x)
case *ast.DropBindingStmt:
return b.buildDropBindPlan(x)
case *ast.SetBindingStmt:
return b.buildSetBindingStatusPlan(x)
case *ast.SplitRegionStmt:
return b.buildSplitRegion(x)
case *ast.DistributeTableStmt:
return b.buildDistributeTable(x)
case *ast.CompactTableStmt:
return b.buildCompactTable(x)
case *ast.RecommendIndexStmt:
return b.buildRecommendIndex(x)
}
return nil, plannererrors.ErrUnsupportedType.GenWithStack("Unsupported type %T", node.Node)
}
func (b *PlanBuilder) buildSetConfig(ctx context.Context, v *ast.SetConfigStmt) (base.Plan, error) {
privErr := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("CONFIG")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ConfigPriv, "", "", "", privErr)
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
if _, ok := v.Value.(*ast.DefaultExpr); ok {
return nil, errors.New("Unknown DEFAULT for SET CONFIG")
}
expr, _, err := b.rewrite(ctx, v.Value, mockTablePlan, nil, true)
return &SetConfig{Name: v.Name, Type: v.Type, Instance: v.Instance, Value: expr}, err
}
func (b *PlanBuilder) buildExecute(ctx context.Context, v *ast.ExecuteStmt) (base.Plan, error) {
vars := make([]expression.Expression, 0, len(v.UsingVars))
for _, expr := range v.UsingVars {
newExpr, _, err := b.rewrite(ctx, expr, nil, nil, true)
if err != nil {
return nil, err
}
vars = append(vars, newExpr)
}
prepStmt, err := GetPreparedStmt(v, b.ctx.GetSessionVars())
if err != nil {
return nil, err
}
exe := &Execute{Name: v.Name, Params: vars, PrepStmt: prepStmt}
if v.BinaryArgs != nil {
exe.Params = v.BinaryArgs.([]expression.Expression)
}
return exe, nil
}
func (b *PlanBuilder) buildDo(ctx context.Context, v *ast.DoStmt) (base.Plan, error) {
var p base.LogicalPlan
dual := logicalop.LogicalTableDual{RowCount: 1}.Init(b.ctx, b.getSelectOffset())
dual.SetSchema(expression.NewSchema())
p = dual
proj := logicalop.LogicalProjection{Exprs: make([]expression.Expression, 0, len(v.Exprs))}.Init(b.ctx, b.getSelectOffset())
proj.SetOutputNames(make([]*types.FieldName, len(v.Exprs)))
schema := expression.NewSchema(make([]*expression.Column, 0, len(v.Exprs))...)
// Since do statement only contain expression list, and it may contain aggFunc, detecting to build the aggMapper firstly.
var (
err error
aggFuncs []*ast.AggregateFuncExpr
totalMap map[*ast.AggregateFuncExpr]int
)
hasAgg := b.detectAggInExprNode(v.Exprs)
needBuildAgg := hasAgg
if hasAgg {
if b.buildingRecursivePartForCTE {
return nil, plannererrors.ErrCTERecursiveForbidsAggregation.GenWithStackByArgs(b.genCTETableNameForError())
}
aggFuncs, totalMap = b.extractAggFuncsInExprs(v.Exprs)
if len(aggFuncs) == 0 {
needBuildAgg = false
}
}
if needBuildAgg {
var aggIndexMap map[int]int
p, aggIndexMap, err = b.buildAggregation(ctx, p, aggFuncs, nil, nil)
if err != nil {
return nil, err
}
for agg, idx := range totalMap {
totalMap[agg] = aggIndexMap[idx]
}
}
for _, astExpr := range v.Exprs {
expr, np, err := b.rewrite(ctx, astExpr, p, totalMap, true)
if err != nil {
return nil, err
}
p = np
proj.Exprs = append(proj.Exprs, expr)
schema.Append(&expression.Column{
UniqueID: b.ctx.GetSessionVars().AllocPlanColumnID(),
RetType: expr.GetType(b.ctx.GetExprCtx().GetEvalCtx()),
})
}
proj.SetChildren(p)
proj.SetSelf(proj)
proj.SetSchema(schema)
proj.CalculateNoDelay = true
return proj, nil
}
func (b *PlanBuilder) buildSet(ctx context.Context, v *ast.SetStmt) (base.Plan, error) {
p := &Set{}
for _, vars := range v.Variables {
if vars.IsGlobal || vars.IsInstance {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or SYSTEM_VARIABLES_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"SYSTEM_VARIABLES_ADMIN"}, false, err)
}
varName := strings.ToLower(vars.Name)
if sem.IsEnabled() && sem.IsInvisibleSysVar(varName) {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("RESTRICTED_VARIABLES_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESTRICTED_VARIABLES_ADMIN"}, false, err)
}
if kerneltype.IsNextGen() && vardef.IsReadOnlyVarInNextGen(varName) {
return nil, variable.ErrNotSupportedInNextGen.GenWithStackByArgs(fmt.Sprintf("setting %s", varName))
}
assign := &expression.VarAssignment{
Name: vars.Name,
IsGlobal: vars.IsGlobal,
IsInstance: vars.IsInstance,
IsSystem: vars.IsSystem,
}
if _, ok := vars.Value.(*ast.DefaultExpr); !ok {
if cn, ok2 := vars.Value.(*ast.ColumnNameExpr); ok2 && cn.Name.Table.L == "" {
// Convert column name expression to string value expression.
char, col := b.ctx.GetSessionVars().GetCharsetInfo()
vars.Value = ast.NewValueExpr(cn.Name.Name.O, char, col)
}
// The mocked plan need one output for the complex cases.
// See the following IF branch.
mockTablePlan := logicalop.LogicalTableDual{RowCount: 1}.Init(b.ctx, b.getSelectOffset())
var err error
var possiblePlan base.LogicalPlan
assign.Expr, possiblePlan, err = b.rewrite(ctx, vars.Value, mockTablePlan, nil, true)
if err != nil {
return nil, err
}
// It's possible that the subquery of the SET_VAR is a complex one so we need to get the result by evaluating the plan.
if _, ok := possiblePlan.(*logicalop.LogicalTableDual); !ok {
physicalPlan, _, err := DoOptimize(ctx, b.ctx, b.optFlag, possiblePlan)
if err != nil {
return nil, err
}
row, err := EvalSubqueryFirstRow(ctx, physicalPlan, b.is, b.ctx)
if err != nil {
return nil, err
}
constant := &expression.Constant{
Value: row[0],
RetType: assign.Expr.GetType(b.ctx.GetExprCtx().GetEvalCtx()),
}
assign.Expr = constant
}
} else {
assign.IsDefault = true
}
if vars.ExtendValue != nil {
assign.ExtendValue = &expression.Constant{
Value: vars.ExtendValue.(*driver.ValueExpr).Datum,
RetType: &vars.ExtendValue.(*driver.ValueExpr).Type,
}
}
p.VarAssigns = append(p.VarAssigns, assign)
}
return p, nil
}
func (b *PlanBuilder) buildDropBindPlan(v *ast.DropBindingStmt) (base.Plan, error) {
var p *SQLBindPlan
if v.OriginNode != nil {
normdOrigSQL, sqlDigestWithDB := bindinfo.NormalizeStmtForBinding(v.OriginNode, b.ctx.GetSessionVars().CurrentDB, false)
p = &SQLBindPlan{
IsGlobal: v.GlobalScope,
SQLBindOp: OpSQLBindDrop,
Details: []*SQLBindOpDetail{{
NormdOrigSQL: normdOrigSQL,
Db: utilparser.GetDefaultDB(v.OriginNode, b.ctx.GetSessionVars().CurrentDB),
SQLDigest: sqlDigestWithDB,
}},
}
if v.HintedNode != nil {
p.Details[0].BindSQL = bindinfo.RestoreDBForBinding(v.HintedNode, b.ctx.GetSessionVars().CurrentDB)
}
} else {
sqlDigests, err := collectStrOrUserVarList(b.ctx, v.SQLDigests)
if err != nil {
return nil, err
}
if len(sqlDigests) == 0 {
return nil, errors.New("sql digest is empty")
}
details := make([]*SQLBindOpDetail, 0, len(sqlDigests))
for _, sqlDigest := range sqlDigests {
details = append(details, &SQLBindOpDetail{SQLDigest: sqlDigest})
}
p = &SQLBindPlan{
SQLBindOp: OpSQLBindDropByDigest,
IsGlobal: v.GlobalScope,
Details: details,
}
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
return p, nil
}
func (b *PlanBuilder) buildSetBindingStatusPlan(v *ast.SetBindingStmt) (base.Plan, error) {
var p *SQLBindPlan
if v.OriginNode != nil {
normdSQL, _ := bindinfo.NormalizeStmtForBinding(v.OriginNode, b.ctx.GetSessionVars().CurrentDB, false)
p = &SQLBindPlan{
SQLBindOp: OpSetBindingStatus,
Details: []*SQLBindOpDetail{{
NormdOrigSQL: normdSQL,
Db: utilparser.GetDefaultDB(v.OriginNode, b.ctx.GetSessionVars().CurrentDB),
}},
}
} else if v.SQLDigest != "" {
p = &SQLBindPlan{
SQLBindOp: OpSetBindingStatusByDigest,
Details: []*SQLBindOpDetail{{
SQLDigest: v.SQLDigest,
}},
}
} else {
return nil, errors.New("sql digest is empty")
}
switch v.BindingStatusType {
case ast.BindingStatusTypeEnabled:
p.Details[0].NewStatus = bindinfo.StatusEnabled
case ast.BindingStatusTypeDisabled:
p.Details[0].NewStatus = bindinfo.StatusDisabled
}
if v.HintedNode != nil {
p.Details[0].BindSQL = bindinfo.RestoreDBForBinding(v.HintedNode, b.ctx.GetSessionVars().CurrentDB)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
return p, nil
}
func checkHintedSQL(sql, charset, collation, db string) error {
p := parser.New()
hintsSet, _, warns, err := hint.ParseHintsSet(p, sql, charset, collation, db)
if err != nil {
return err
}
hintsStr, err := hintsSet.Restore()
if err != nil {
return err
}
// For `create global binding for select * from t using select * from t`, we allow it though hintsStr is empty.
// For `create global binding for select * from t using select /*+ non_exist_hint() */ * from t`,
// the hint is totally invalid, we escalate warning to error.
if hintsStr == "" && len(warns) > 0 {
return warns[0]
}
return nil
}
func fetchRecordFromClusterStmtSummary(sctx base.PlanContext, planDigest string) (schema, query, planHint, charset, collation string, err error) {
ctx := kv.WithInternalSourceType(context.Background(), kv.InternalTxnBindInfo)
exec := sctx.GetSQLExecutor()
prefix := "cluster_"
if intest.InTest {
prefix = ""
}
fields := "stmt_type, schema_name, digest_text, sample_user, prepared, query_sample_text, charset, collation, plan_hint, plan_digest"
sql := fmt.Sprintf("select %s from information_schema.%stidb_statements_stats where plan_digest = '%s' union distinct ", fields, prefix, planDigest) +
fmt.Sprintf("select %s from information_schema.%sstatements_summary_history where plan_digest = '%s' ", fields, prefix, planDigest) +
"order by length(plan_digest) desc"
rs, err := exec.ExecuteInternal(ctx, sql)
if err != nil {
return "", "", "", "", "", err
}
if rs == nil {
return "", "", "", "", "",
errors.New("can't find any records for '" + planDigest + "' in statement summary")
}
var rows []chunk.Row
defer terror.Call(rs.Close)
if rows, err = sqlexec.DrainRecordSet(ctx, rs, 8); err != nil {
return "", "", "", "", "", err
}
for _, row := range rows {
user := row.GetString(3)
stmtType := row.GetString(0)
if user != "" && (stmtType == "Select" || stmtType == "Delete" || stmtType == "Update" || stmtType == "Insert" || stmtType == "Replace") {
// Empty auth users means that it is an internal queries.
schema = row.GetString(1)
query = row.GetString(5)
planHint = row.GetString(8)
charset = row.GetString(6)
collation = row.GetString(7)
// If it is SQL command prepare / execute, we should remove the arguments
// If it is binary protocol prepare / execute, ssbd.normalizedSQL should be same as ssElement.sampleSQL.
if row.GetInt64(4) == 1 {
if idx := strings.LastIndex(query, "[arguments:"); idx != -1 {
query = query[:idx]
}
}
return
}
}
return
}
func collectStrOrUserVarList(ctx base.PlanContext, list []*ast.StringOrUserVar) ([]string, error) {
result := make([]string, 0, len(list))
for _, single := range list {
var str string
if single.UserVar != nil {
val, ok := ctx.GetSessionVars().GetUserVarVal(strings.ToLower(single.UserVar.Name))
if !ok {
return nil, errors.New("can't find specified user variable: " + single.UserVar.Name)
}
var err error
str, err = val.ToString()
if err != nil {
return nil, err
}
} else {
str = single.StringLit
}
split := strings.Split(str, ",")
for _, single := range split {
trimmed := strings.TrimSpace(single)
if len(trimmed) > 0 {
result = append(result, trimmed)
}
}
}
return result, nil
}
// constructSQLBindOPFromPlanDigest tries to construct a SQLBindOpDetail from plan digest by fetching the corresponding
// record from cluster_statements_summary or cluster_statements_summary_history.
// If it fails to construct the SQLBindOpDetail for any reason, it will return (nil, error).
// If the plan digest corresponds to the same SQL digest as another one in handledSQLDigests, it will append a warning
// then return (nil, nil).
func constructSQLBindOPFromPlanDigest(
ctx base.PlanContext,
planDigest string,
handledSQLDigests map[string]struct{},
) (
*SQLBindOpDetail,
error,
) {
// The warnings will be broken in fetchRecordFromClusterStmtSummary(), so we need to save and restore it to make the
// warnings for repeated SQL Digest work.
warnings := ctx.GetSessionVars().StmtCtx.GetWarnings()
schema, query, planHint, charset, collation, err := fetchRecordFromClusterStmtSummary(ctx, planDigest)
if err != nil {
return nil, err
}
if query == "" {
return nil, errors.New("can't find any plans for '" + planDigest + "'")
}
// Check if the SQL is truncated (ends with "(len:<num>)" pattern)
if query[len(query)-1] == ')' {
match, err2 := regexp.MatchString(`\(len:\d+\)$`, query)
if match || err2 != nil {
return nil, errors.NewNoStackErrorf("binding failed: SQL query is truncated due to tidb_stmt_summary_max_sql_length limit. "+
"Please increase tidb_stmt_summary_max_sql_length. Plan Digest: %v", planDigest)
}
}
ctx.GetSessionVars().StmtCtx.SetWarnings(warnings)
parser4binding := parser.New()
originNode, err := parser4binding.ParseOneStmt(query, charset, collation)
if err != nil {
return nil, errors.NewNoStackErrorf("binding failed: %v. Plan Digest: %v", err, planDigest)
}
complete, reason := hint.CheckBindingFromHistoryComplete(originNode, planHint)
bindSQL := bindinfo.GenerateBindingSQL(originNode, planHint, schema)
var hintNode ast.StmtNode
hintNode, err = parser4binding.ParseOneStmt(bindSQL, charset, collation)
if err != nil {
return nil, errors.NewNoStackErrorf("binding failed: %v. Plan Digest: %v", err, planDigest)
}
bindSQL = bindinfo.RestoreDBForBinding(hintNode, schema)
db := utilparser.GetDefaultDB(originNode, schema)
normdOrigSQL, sqlDigestWithDBStr := bindinfo.NormalizeStmtForBinding(originNode, schema, false)
if _, ok := handledSQLDigests[sqlDigestWithDBStr]; ok {
ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError(
planDigest + " is ignored because it corresponds to the same SQL digest as another Plan Digest",
))
return nil, nil
}
handledSQLDigests[sqlDigestWithDBStr] = struct{}{}
if !complete {
ctx.GetSessionVars().StmtCtx.AppendWarning(
errors.NewNoStackErrorf("%v. Plan Digest: %v", reason, planDigest),
)
}
op := &SQLBindOpDetail{
NormdOrigSQL: normdOrigSQL,
BindSQL: bindSQL,
BindStmt: hintNode,
Db: db,
Charset: charset,
Collation: collation,
Source: bindinfo.SourceHistory,
SQLDigest: sqlDigestWithDBStr,
PlanDigest: planDigest,
}
return op, nil
}
func (b *PlanBuilder) buildCreateBindPlanFromPlanDigest(v *ast.CreateBindingStmt) (base.Plan, error) {
planDigests, err := collectStrOrUserVarList(b.ctx, v.PlanDigests)
if err != nil {
return nil, err
}
if len(planDigests) == 0 {
return nil, errors.New("plan digest is empty")
}
handledSQLDigests := make(map[string]struct{}, len(planDigests))
opDetails := make([]*SQLBindOpDetail, 0, len(planDigests))
for _, planDigest := range planDigests {
op, err2 := constructSQLBindOPFromPlanDigest(b.ctx, planDigest, handledSQLDigests)
if err2 != nil {
return nil, err2
}
if op == nil {
continue
}
opDetails = append(opDetails, op)
}
p := &SQLBindPlan{
IsGlobal: v.GlobalScope,
SQLBindOp: OpSQLBindCreate,
Details: opDetails,
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
return p, nil
}
func (b *PlanBuilder) buildCreateBindPlan(v *ast.CreateBindingStmt) (base.Plan, error) {
if v.OriginNode == nil {
return b.buildCreateBindPlanFromPlanDigest(v)
}
charSet, collation := b.ctx.GetSessionVars().GetCharsetInfo()
// Because we use HintedNode.Restore instead of HintedNode.Text, so we need do some check here
// For example, if HintedNode.Text is `select /*+ non_exist_hint() */ * from t` and the current DB is `test`,
// the HintedNode.Restore will be `select * from test . t`.
// In other words, illegal hints will be deleted during restore. We can't check hinted SQL after restore.
// So we need check here.
if err := checkHintedSQL(v.HintedNode.Text(), charSet, collation, b.ctx.GetSessionVars().CurrentDB); err != nil {
return nil, err
}
bindSQL := bindinfo.RestoreDBForBinding(v.HintedNode, b.ctx.GetSessionVars().CurrentDB)
db := utilparser.GetDefaultDB(v.OriginNode, b.ctx.GetSessionVars().CurrentDB)
normdOrigSQL, sqlDigestWithDB := bindinfo.NormalizeStmtForBinding(v.OriginNode, b.ctx.GetSessionVars().CurrentDB, false)
p := &SQLBindPlan{
IsGlobal: v.GlobalScope,
SQLBindOp: OpSQLBindCreate,
Details: []*SQLBindOpDetail{{
NormdOrigSQL: normdOrigSQL,
BindSQL: bindSQL,
BindStmt: v.HintedNode,
Db: db,
Charset: charSet,
Collation: collation,
Source: bindinfo.SourceManual,
SQLDigest: sqlDigestWithDB,
}},
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
return p, nil
}
// detectAggInExprNode detects an aggregate function in its exprs.
func (*PlanBuilder) detectAggInExprNode(exprs []ast.ExprNode) bool {
return slices.ContainsFunc(exprs, ast.HasAggFlag)
}
// detectSelectAgg detects an aggregate function or GROUP BY clause.
func (*PlanBuilder) detectSelectAgg(sel *ast.SelectStmt) bool {
if sel.GroupBy != nil {
return true
}
for _, f := range sel.Fields.Fields {
if f.WildCard != nil {
continue
}
if ast.HasAggFlag(f.Expr) {
return true
}
}
if sel.Having != nil {
if ast.HasAggFlag(sel.Having.Expr) {
return true
}
}
if sel.OrderBy != nil {
for _, item := range sel.OrderBy.Items {
if ast.HasAggFlag(item.Expr) {
return true
}
}
}
return false
}
func (*PlanBuilder) detectSelectWindow(sel *ast.SelectStmt) bool {
for _, f := range sel.Fields.Fields {
if ast.HasWindowFlag(f.Expr) {
return true
}
}
if sel.OrderBy != nil {
for _, item := range sel.OrderBy.Items {
if ast.HasWindowFlag(item.Expr) {
return true
}
}
}
return false
}
func getPathByIndexName(paths []*util.AccessPath, idxName ast.CIStr, tblInfo *model.TableInfo) *util.AccessPath {
var indexPrefixPath *util.AccessPath
prefixMatches := 0
for _, path := range paths {
// Only accept tikv's primary key table path.
if path.IsTiKVTablePath() && isPrimaryIndex(idxName) && tblInfo.HasClusteredIndex() {
return path
}
// If it's not a tikv table path and the index is nil, it could not be any index path.
if path.Index == nil {
continue
}
if path.Index.Name.L == idxName.L {
return path
}
if strings.HasPrefix(path.Index.Name.L, idxName.L) {
indexPrefixPath = path
prefixMatches++
}
}
// Return only unique prefix matches
if prefixMatches == 1 {
return indexPrefixPath
}
return nil
}
func isPrimaryIndex(indexName ast.CIStr) bool {
return indexName.L == "primary"
}
func genTiFlashPath(tblInfo *model.TableInfo) *util.AccessPath {
tiFlashPath := &util.AccessPath{StoreType: kv.TiFlash}
fillContentForTablePath(tiFlashPath, tblInfo)
return tiFlashPath
}
func fillContentForTablePath(tablePath *util.AccessPath, tblInfo *model.TableInfo) {
if tblInfo.IsCommonHandle {
tablePath.IsCommonHandlePath = true
for _, index := range tblInfo.Indices {
if index.Primary {
tablePath.Index = index
break
}
}
} else {
tablePath.IsIntHandlePath = true
}
}
// isForUpdateReadSelectLock checks if the lock type need to use forUpdateRead
func isForUpdateReadSelectLock(lock *ast.SelectLockInfo) bool {
if lock == nil {
return false
}
return lock.LockType == ast.SelectLockForUpdate ||
lock.LockType == ast.SelectLockForUpdateNoWait ||
lock.LockType == ast.SelectLockForUpdateWaitN
}
func isTiKVIndexByName(idxName ast.CIStr, indexInfo *model.IndexInfo, tblInfo *model.TableInfo) bool {
// when the PKIsHandle of table is true, the primary key is not in the indices list.
if idxName.L == "primary" && tblInfo.PKIsHandle {
return true
}
return indexInfo != nil && !indexInfo.IsColumnarIndex()
}
func checkIndexLookUpPushDownSupported(ctx base.PlanContext, tblInfo *model.TableInfo, index *model.IndexInfo, suppressWarning bool) bool {
unSupportedReason := ""
sessionVars := ctx.GetSessionVars()
if tblInfo.IsCommonHandle && tblInfo.CommonHandleVersion < 1 {
unSupportedReason = "common handle table with old encoding version is not supported"
} else if index.Global {
unSupportedReason = "the global index in partition table is not supported"
} else if tblInfo.TempTableType != model.TempTableNone {
unSupportedReason = "temporary table is not supported"
} else if tblInfo.TableCacheStatusType != model.TableCacheStatusDisable {
unSupportedReason = "cached table is not supported"
} else if index.MVIndex {
unSupportedReason = "multi-valued index is not supported"
} else if !sessionVars.IsIsolation(ast.RepeatableRead) {
unSupportedReason = "transaction isolation level is not REPEATABLE-READ"
} else if sessionVars.GetReplicaRead() != kv.ReplicaReadLeader {
unSupportedReason = "only leader read is supported"
} else if sessionVars.TxnCtx.IsStaleness {
unSupportedReason = "stale read is not supported"
} else if sessionVars.SnapshotTS != 0 {
unSupportedReason = "historical read is not supported"
}
if unSupportedReason != "" {
if !suppressWarning {
ctx.GetSessionVars().StmtCtx.SetHintWarning(fmt.Sprintf("hint INDEX_LOOKUP_PUSHDOWN is inapplicable, %s", unSupportedReason))
}
return false
}
return true
}
func checkAutoForceIndexLookUpPushDown(ctx base.PlanContext, tblInfo *model.TableInfo, index *model.IndexInfo) bool {
policy := ctx.GetSessionVars().IndexLookUpPushDownPolicy
switch policy {
case vardef.IndexLookUpPushDownPolicyForce:
case vardef.IndexLookUpPushDownPolicyAffinityForce:
if tblInfo.Affinity == nil {
// No affinity info, should not auto push down the index look up.
return false
}
default:
return false
}
return checkIndexLookUpPushDownSupported(ctx, tblInfo, index, true)
}
func getPossibleAccessPaths(ctx base.PlanContext, tableHints *hint.PlanHints, indexHints []*ast.IndexHint, tbl table.Table, dbName, tblName ast.CIStr, check bool, hasFlagPartitionProcessor bool) ([]*util.AccessPath, error) {
tblInfo := tbl.Meta()
publicPaths := make([]*util.AccessPath, 0, len(tblInfo.Indices)+2)
tp := kv.TiKV
if tbl.Type().IsClusterTable() {
tp = kv.TiDB
}
tablePath := &util.AccessPath{StoreType: tp}
fillContentForTablePath(tablePath, tblInfo)
publicPaths = append(publicPaths, tablePath)
// consider hypo TiFlash replicas
isHypoTiFlashReplica := false
if ctx.GetSessionVars().StmtCtx.InExplainStmt && ctx.GetSessionVars().HypoTiFlashReplicas != nil {
hypoReplicas := ctx.GetSessionVars().HypoTiFlashReplicas
originalTableName := tblInfo.Name.L
if hypoReplicas[dbName.L] != nil {
if _, ok := hypoReplicas[dbName.L][originalTableName]; ok {
isHypoTiFlashReplica = true
}
}
}
if tblInfo.TiFlashReplica == nil {
if isHypoTiFlashReplica {
publicPaths = append(publicPaths, genTiFlashPath(tblInfo))
} else {
ctx.GetSessionVars().RaiseWarningWhenMPPEnforced("MPP mode may be blocked because there aren't tiflash replicas of table `" + tblInfo.Name.O + "`.")
}
} else if !tblInfo.TiFlashReplica.Available {
if isHypoTiFlashReplica {
publicPaths = append(publicPaths, genTiFlashPath(tblInfo))
} else {
ctx.GetSessionVars().RaiseWarningWhenMPPEnforced("MPP mode may be blocked because tiflash replicas of table `" + tblInfo.Name.O + "` not ready.")
}
} else {
publicPaths = append(publicPaths, genTiFlashPath(tblInfo))
}
optimizerUseInvisibleIndexes := ctx.GetSessionVars().OptimizerUseInvisibleIndexes
check = check || ctx.GetSessionVars().IsIsolation(ast.ReadCommitted)
check = check && ctx.GetSessionVars().ConnectionID > 0
var latestIndexes map[int64]*model.IndexInfo
var err error
// Inverted Index can not be used as access path index.
invertedIndexes := make(map[string]struct{})
// When NO_INDEX_LOOKUP_PUSHDOWN hint is specified, we should set `forceNoIndexLookUpPushDown = true` to avoid
// using index look up push down even if other hint or system variable `tidb_index_lookup_pushdown_policy`
// tries to enable it.
forceNoIndexLookUpPushDown := false
if len(tableHints.NoIndexLookUpPushDown) > 0 {
for _, h := range tableHints.NoIndexLookUpPushDown {
if h.Match(&hint.HintedTable{
DBName: dbName,
TblName: tblName,
}) {
h.Matched = true
forceNoIndexLookUpPushDown = true
break
}
}
}
for _, index := range tblInfo.Indices {
if index.State == model.StatePublic {
// Filter out invisible index, because they are not visible for optimizer
if !optimizerUseInvisibleIndexes && index.Invisible {
continue
}
if tblInfo.IsCommonHandle && index.Primary {
continue
}
if check && latestIndexes == nil {
latestIndexes, check, err = domainmisc.GetLatestIndexInfo(ctx, tblInfo.ID, 0)
if err != nil {
return nil, err
}
}
if check {
if latestIndex, ok := latestIndexes[index.ID]; !ok || latestIndex.State != model.StatePublic {
continue
}
}
if index.InvertedInfo != nil {
invertedIndexes[index.Name.L] = struct{}{}
continue
}
if index.IsColumnarIndex() {
// Because the value of `TiFlashReplica.Available` changes as the user modify replica, it is not ideal if the state of index changes accordingly.
// So the current way to use the columnar indexes is to require the TiFlash Replica to be available.
if !tblInfo.TiFlashReplica.Available {
continue
}
path := genTiFlashPath(tblInfo)
path.Index = index
publicPaths = append(publicPaths, path)
continue
}
path := &util.AccessPath{Index: index}
if !forceNoIndexLookUpPushDown && checkAutoForceIndexLookUpPushDown(ctx, tblInfo, index) {
path.IndexLookUpPushDownBy = util.IndexLookUpPushDownBySysVar
}
publicPaths = append(publicPaths, path)
}
}
// consider hypo-indexes
hypoIndexes := ctx.GetSessionVars().HypoIndexes // session level hypo-indexes
if ctx.GetSessionVars().StmtCtx.InExplainStmt && hypoIndexes != nil {
originalTableName := tblInfo.Name.L
if hypoIndexes[dbName.L] != nil && hypoIndexes[dbName.L][originalTableName] != nil {
for _, index := range hypoIndexes[dbName.L][originalTableName] {
publicPaths = append(publicPaths, &util.AccessPath{Index: index})
}
}
}
if len(ctx.GetSessionVars().StmtCtx.HintedHypoIndexes) > 0 { // statement-level hypo-indexes from hints
if !ctx.GetSessionVars().StmtCtx.InExplainStmt {
ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError("can only use HYPO_INDEX hint in explain-statements"))
} else {
hintedHypoIndexes := ctx.GetSessionVars().StmtCtx.HintedHypoIndexes
originalTableName := tblInfo.Name.L
if hintedHypoIndexes[dbName.L] != nil && hintedHypoIndexes[dbName.L][originalTableName] != nil {
for _, index := range hintedHypoIndexes[dbName.L][originalTableName] {
publicPaths = append(publicPaths, &util.AccessPath{Index: index})
}
}
}
}
hasScanHint, hasUseOrForce := false, false
available := make([]*util.AccessPath, 0, len(publicPaths))
ignored := make([]*util.AccessPath, 0, len(publicPaths))
// Extract comment-style index hint like /*+ INDEX(t, idx1, idx2) */.
indexHintsLen := len(indexHints)
var indexLookUpPushDownHints map[int]struct{}
if tableHints != nil {
for i, hint := range tableHints.IndexHintList {
if hint.Match(dbName, tblName) {
indexHints = append(indexHints, hint.IndexHint)
tableHints.IndexHintList[i].Matched = true
if hint.ShouldPushDownIndexLookUp() {
if indexLookUpPushDownHints == nil {
indexLookUpPushDownHints = make(map[int]struct{}, 1)
}
indexLookUpPushDownHints[len(indexHints)-1] = struct{}{}
}
}
}
}
_, isolationReadEnginesHasTiKV := ctx.GetSessionVars().GetIsolationReadEngines()[kv.TiKV]
for i, hint := range indexHints {
if hint.HintScope != ast.HintForScan {
continue
}
hasScanHint = true
// It is syntactically valid to omit index_list for USE INDEX, which means "use no indexes".
// Omitting index_list for FORCE INDEX or IGNORE INDEX is a syntax error.
// See https://dev.mysql.com/doc/refman/8.0/en/index-hints.html.
if !isolationReadEnginesHasTiKV && hint.IndexNames == nil {
continue
}
if hint.IndexNames == nil && hint.HintType != ast.HintIgnore {
if path := getTablePath(publicPaths); path != nil {
hasUseOrForce = true
path.Forced = true
available = append(available, path)
}
}
for _, idxName := range hint.IndexNames {
if _, ok := invertedIndexes[idxName.L]; ok {
continue
}
path := getPathByIndexName(publicPaths, idxName, tblInfo)
if path == nil {
err := plannererrors.ErrKeyDoesNotExist.FastGenByArgs(idxName, tblInfo.Name)
// if hint is from comment-style sql hints, we should throw a warning instead of error.
if i < indexHintsLen {
return nil, err
}
ctx.GetSessionVars().StmtCtx.AppendWarning(err)
continue
}
if hint.HintType == ast.HintIgnore {
// Collect all the ignored index hints.
ignored = append(ignored, path)
continue
}
if isTiKVIndexByName(idxName, path.Index, tblInfo) && !isolationReadEnginesHasTiKV {
engineVals, _ := ctx.GetSessionVars().GetSystemVar(vardef.TiDBIsolationReadEngines)
err := fmt.Errorf("TiDB doesn't support index '%v' in the isolation read engines(value: '%v')", idxName, engineVals)
if i < indexHintsLen {
return nil, err
}
ctx.GetSessionVars().StmtCtx.AppendWarning(err)
continue
}
// Currently we don't distinguish between "FORCE" and "USE" because
// our cost estimation is not reliable.
hasUseOrForce = true
path.Forced = true
if hint.HintType == ast.HintOrderIndex {
path.ForceKeepOrder = true
}
if hint.HintType == ast.HintNoOrderIndex {
path.ForceNoKeepOrder = true
}
if i >= indexHintsLen {
// Currently we only support to hint the index look up push down for comment-style sql hints.
// So only i >= indexHintsLen may have the hints here.
if _, ok := indexLookUpPushDownHints[i]; ok {
if forceNoIndexLookUpPushDown {
ctx.GetSessionVars().StmtCtx.SetHintWarning(
"hint INDEX_LOOKUP_PUSHDOWN cannot be inapplicable, NO_INDEX_LOOKUP_PUSHDOWN is specified",
)
continue
}
if !checkIndexLookUpPushDownSupported(ctx, tblInfo, path.Index, false) {
continue
}
path.IndexLookUpPushDownBy = util.IndexLookUpPushDownByHint
}
}
available = append(available, path)
}
}
if !hasScanHint || !hasUseOrForce {
available = publicPaths
}
available = removeIgnoredPaths(available, ignored, tblInfo)
// global index must not use partition pruning optimization, as LogicalPartitionAll not suitable for global index.
// ignore global index if flagPartitionProcessor exists.
if hasFlagPartitionProcessor {
available = removeGlobalIndexPaths(available)
}
// If we have got "FORCE" or "USE" index hint but got no available index,
// we have to use table scan.
if len(available) == 0 {
available = append(available, tablePath)
}
// If all available paths are Multi-Valued Index, it's possible that the only multi-valued index is inapplicable,
// so that the table paths are still added here to avoid failing to find any physical plan.
allMVIIndexPath := true
for _, availablePath := range available {
if !isMVIndexPath(availablePath) {
allMVIIndexPath = false
break
}
}
if allMVIIndexPath {
available = append(available, tablePath)
}
return available, nil
}
func removeIgnoredPaths(paths, ignoredPaths []*util.AccessPath, tblInfo *model.TableInfo) []*util.AccessPath {
if len(ignoredPaths) == 0 {
return paths
}
remainedPaths := make([]*util.AccessPath, 0, len(paths))
for _, path := range paths {
if path.IsTiKVTablePath() || path.IsTiFlashSimpleTablePath() || getPathByIndexName(ignoredPaths, path.Index.Name, tblInfo) == nil {
remainedPaths = append(remainedPaths, path)
}
}
return remainedPaths
}
func removeGlobalIndexPaths(paths []*util.AccessPath) []*util.AccessPath {
i := 0
for _, path := range paths {
if path.Index != nil && path.Index.Global {
continue
}
paths[i] = path
i++
}
return paths[:i]
}
func (b *PlanBuilder) buildSelectLock(src base.LogicalPlan, lock *ast.SelectLockInfo) (*logicalop.LogicalLock, error) {
var tblID2PhysTblIDCol map[int64]*expression.Column
if len(b.partitionedTable) > 0 {
tblID2PhysTblIDCol = make(map[int64]*expression.Column)
// If a chunk row is read from a partitioned table, which partition the row
// comes from is unknown. With the existence of Join, the situation could be
// even worse: SelectLock have to know the `pid` to construct the lock key.
// To solve the problem, an extra `pid` column is added to the schema, and the
// DataSource need to return the `pid` information in the chunk row.
// For dynamic prune mode, it is filled in from the tableID in the key by storage.
// For static prune mode it is also filled in from the tableID in the key by storage.
// since it would otherwise be lost in the PartitionUnion executor.
setExtraPhysTblIDColsOnDataSource(src, tblID2PhysTblIDCol)
}
selectLock := logicalop.LogicalLock{
Lock: lock,
TblID2Handle: b.handleHelper.tailMap(),
TblID2PhysTblIDCol: tblID2PhysTblIDCol,
}.Init(b.ctx)
selectLock.SetChildren(src)
return selectLock, nil
}
func setExtraPhysTblIDColsOnDataSource(p base.LogicalPlan, tblID2PhysTblIDCol map[int64]*expression.Column) {
switch ds := p.(type) {
case *logicalop.DataSource:
if ds.TableInfo.GetPartitionInfo() == nil {
return
}
tblID2PhysTblIDCol[ds.TableInfo.ID] = addExtraPhysTblIDColumn4DS(ds)
default:
for _, child := range p.Children() {
setExtraPhysTblIDColsOnDataSource(child, tblID2PhysTblIDCol)
}
}
}
func (b *PlanBuilder) buildPrepare(x *ast.PrepareStmt) base.Plan {
p := &Prepare{
Name: x.Name,
}
if x.SQLVar != nil {
if v, ok := b.ctx.GetSessionVars().GetUserVarVal(strings.ToLower(x.SQLVar.Name)); ok {
var err error
p.SQLText, err = v.ToString()
if err != nil {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(err)
p.SQLText = "NULL"
}
} else {
p.SQLText = "NULL"
}
} else {
p.SQLText = x.SQLText
}
return p
}
func (b *PlanBuilder) buildAdmin(ctx context.Context, as *ast.AdminStmt) (base.Plan, error) {
var ret base.Plan
var err error
switch as.Tp {
case ast.AdminCheckTable, ast.AdminCheckIndex:
ret, err = b.buildAdminCheckTable(ctx, as)
if err != nil {
return ret, err
}
case ast.AdminRecoverIndex:
tnW := b.resolveCtx.GetTableName(as.Tables[0])
p := &RecoverIndex{Table: tnW, IndexName: as.Index}
p.SetSchemaAndNames(buildRecoverIndexFields())
ret = p
case ast.AdminCleanupIndex:
tnW := b.resolveCtx.GetTableName(as.Tables[0])
p := &CleanupIndex{Table: tnW, IndexName: as.Index}
p.SetSchemaAndNames(buildCleanupIndexFields())
ret = p
case ast.AdminChecksumTable:
tnWs := make([]*resolve.TableNameW, 0, len(as.Tables))
for _, tn := range as.Tables {
tnWs = append(tnWs, b.resolveCtx.GetTableName(tn))
}
p := &ChecksumTable{Tables: tnWs}
p.SetSchemaAndNames(buildChecksumTableSchema())
ret = p
case ast.AdminShowNextRowID:
p := &ShowNextRowID{TableName: as.Tables[0]}
p.SetSchemaAndNames(buildShowNextRowID())
ret = p
case ast.AdminShowDDL:
p := &ShowDDL{}
p.SetSchemaAndNames(buildShowDDLFields())
ret = p
case ast.AdminShowDDLJobs:
p := logicalop.LogicalShowDDLJobs{JobNumber: as.JobNumber}.Init(b.ctx)
p.SetSchemaAndNames(buildShowDDLJobsFields())
for _, col := range p.Schema().Columns {
col.UniqueID = b.ctx.GetSessionVars().AllocPlanColumnID()
}
ret = p
if as.Where != nil {
ret, err = b.buildSelection(ctx, p, as.Where, nil)
if err != nil {
return nil, err
}
}
case ast.AdminCancelDDLJobs:
p := &CancelDDLJobs{JobIDs: as.JobIDs}
p.SetSchemaAndNames(buildCancelDDLJobsFields())
ret = p
case ast.AdminPauseDDLJobs:
p := &PauseDDLJobs{JobIDs: as.JobIDs}
p.SetSchemaAndNames(buildPauseDDLJobsFields())
ret = p
case ast.AdminResumeDDLJobs:
p := &ResumeDDLJobs{JobIDs: as.JobIDs}
p.SetSchemaAndNames(buildResumeDDLJobsFields())
ret = p
case ast.AdminCheckIndexRange:
schema, names, err := b.buildCheckIndexSchema(as.Tables[0], as.Index)
if err != nil {
return nil, err
}
p := &CheckIndexRange{Table: as.Tables[0], IndexName: as.Index, HandleRanges: as.HandleRanges}
p.SetSchemaAndNames(schema, names)
ret = p
case ast.AdminShowDDLJobQueries:
p := &ShowDDLJobQueries{JobIDs: as.JobIDs}
p.SetSchemaAndNames(buildShowDDLJobQueriesFields())
ret = p
case ast.AdminShowDDLJobQueriesWithRange:
p := &ShowDDLJobQueriesWithRange{Limit: as.LimitSimple.Count, Offset: as.LimitSimple.Offset}
p.SetSchemaAndNames(buildShowDDLJobQueriesWithRangeFields())
ret = p
case ast.AdminShowSlow:
p := &ShowSlow{ShowSlow: as.ShowSlow}
p.SetSchemaAndNames(buildShowSlowSchema())
ret = p
case ast.AdminReloadExprPushdownBlacklist:
return &ReloadExprPushdownBlacklist{}, nil
case ast.AdminReloadOptRuleBlacklist:
return &ReloadOptRuleBlacklist{}, nil
case ast.AdminPluginEnable:
return &AdminPlugins{Action: Enable, Plugins: as.Plugins}, nil
case ast.AdminPluginDisable:
return &AdminPlugins{Action: Disable, Plugins: as.Plugins}, nil
case ast.AdminFlushBindings:
return &SQLBindPlan{SQLBindOp: OpFlushBindings}, nil
case ast.AdminCaptureBindings:
return nil, errors.Errorf("Auto Capture is not supported")
case ast.AdminEvolveBindings:
return nil, errors.Errorf("Cannot enable baseline evolution feature, it is not generally available now")
case ast.AdminReloadBindings:
return &SQLBindPlan{SQLBindOp: OpReloadBindings}, nil
case ast.AdminReloadClusterBindings:
return &SQLBindPlan{SQLBindOp: OpReloadClusterBindings}, nil
case ast.AdminReloadStatistics:
return &Simple{Statement: as, ResolveCtx: b.resolveCtx}, nil
case ast.AdminFlushPlanCache:
return &Simple{Statement: as, ResolveCtx: b.resolveCtx}, nil
case ast.AdminSetBDRRole, ast.AdminUnsetBDRRole:
ret = &Simple{Statement: as, ResolveCtx: b.resolveCtx}
case ast.AdminShowBDRRole:
p := &AdminShowBDRRole{}
p.SetSchemaAndNames(buildAdminShowBDRRoleFields())
ret = p
case ast.AdminAlterDDLJob:
ret, err = b.buildAdminAlterDDLJob(ctx, as)
if err != nil {
return nil, err
}
case ast.AdminWorkloadRepoCreate:
ret = &WorkloadRepoCreate{}
default:
return nil, plannererrors.ErrUnsupportedType.GenWithStack("Unsupported ast.AdminStmt(%T) for buildAdmin", as)
}
// Admin command can only be executed by administrator.
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
return ret, nil
}
func (b *PlanBuilder) buildPhysicalIndexLookUpReader(_ context.Context, dbName ast.CIStr, tbl table.Table, idx *model.IndexInfo) (base.Plan, error) {
tblInfo := tbl.Meta()
physicalID, isPartition := getPhysicalID(tbl, idx.Global)
fullExprCols, _, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), dbName, tblInfo)
if err != nil {
return nil, err
}
extraInfo, extraCol, hasExtraCol := tryGetPkExtraColumn(b.ctx.GetSessionVars(), tblInfo)
pkHandleInfo, pkHandleCol, hasPkIsHandle := tryGetPkHandleCol(tblInfo, fullExprCols)
commonInfos, commonCols, hasCommonCols := tryGetCommonHandleCols(tbl, fullExprCols)
idxColInfos := getIndexColumnInfos(tblInfo, idx)
idxColSchema := getIndexColsSchema(tblInfo, idx, fullExprCols)
idxCols, idxColLens := util.IndexInfo2PrefixCols(idxColInfos, idxColSchema.Columns, idx)
pseudoHistColl := statistics.PseudoHistColl(physicalID, false)
is := physicalop.PhysicalIndexScan{
Table: tblInfo,
TableAsName: &tblInfo.Name,
DBName: dbName,
Columns: idxColInfos,
Index: idx,
IdxCols: idxCols,
IdxColLens: idxColLens,
DataSourceSchema: idxColSchema.Clone(),
Ranges: ranger.FullRange(),
PhysicalTableID: physicalID,
IsPartition: isPartition,
TblColHists: &pseudoHistColl,
}.Init(b.ctx, b.getSelectOffset())
// There is no alternative plan choices, so just use pseudo stats to avoid panic.
is.SetStats(&property.StatsInfo{HistColl: &pseudoHistColl})
if hasCommonCols {
for _, c := range commonInfos {
is.Columns = append(is.Columns, c.ColumnInfo)
}
}
is.InitSchema(append(is.IdxCols, commonCols...), true)
// It's double read case.
ts := physicalop.PhysicalTableScan{
Columns: idxColInfos,
Table: tblInfo,
TableAsName: &tblInfo.Name,
DBName: dbName,
PhysicalTableID: physicalID,
TblColHists: &pseudoHistColl,
}.Init(b.ctx, b.getSelectOffset())
ts.SetIsPartition(isPartition)
ts.SetSchema(idxColSchema)
ts.Columns = physicalop.ExpandVirtualColumn(ts.Columns, ts.Schema(), ts.Table.Columns)
switch {
case hasExtraCol:
ts.Columns = append(ts.Columns, extraInfo)
ts.Schema().Append(extraCol)
ts.HandleIdx = []int{len(ts.Columns) - 1}
case hasPkIsHandle:
ts.Columns = append(ts.Columns, pkHandleInfo)
ts.Schema().Append(pkHandleCol)
ts.HandleIdx = []int{len(ts.Columns) - 1}
case hasCommonCols:
ts.HandleIdx = make([]int, 0, len(commonCols))
for pkOffset, cInfo := range commonInfos {
found := false
for i, c := range ts.Columns {
if c.ID == cInfo.ID {
found = true
ts.HandleIdx = append(ts.HandleIdx, i)
break
}
}
if !found {
ts.Columns = append(ts.Columns, cInfo.ColumnInfo)
ts.Schema().Append(commonCols[pkOffset])
ts.HandleIdx = append(ts.HandleIdx, len(ts.Columns)-1)
}
}
}
if is.Index.Global {
tmpColumns, tmpSchema, _ := physicalop.AddExtraPhysTblIDColumn(b.ctx, ts.Columns, ts.Schema())
ts.Columns = tmpColumns
ts.SetSchema(tmpSchema)
}
cop := &physicalop.CopTask{
IndexPlan: is,
TablePlan: ts,
TblColHists: is.StatsInfo().HistColl,
ExtraHandleCol: extraCol,
CommonHandleCols: commonCols,
}
rootT := cop.ConvertToRootTask(b.ctx).(*physicalop.RootTask)
if err := rootT.GetPlan().ResolveIndices(); err != nil {
return nil, err
}
return rootT.GetPlan(), nil
}
func getIndexColumnInfos(tblInfo *model.TableInfo, idx *model.IndexInfo) []*model.ColumnInfo {
ret := make([]*model.ColumnInfo, len(idx.Columns))
for i, idxCol := range idx.Columns {
ret[i] = tblInfo.Columns[idxCol.Offset]
}
return ret
}
func getIndexColsSchema(tblInfo *model.TableInfo, idx *model.IndexInfo, allColSchema *expression.Schema) *expression.Schema {
schema := expression.NewSchema(make([]*expression.Column, 0, len(idx.Columns))...)
for _, idxCol := range idx.Columns {
for i, colInfo := range tblInfo.Columns {
if colInfo.Name.L == idxCol.Name.L {
schema.Append(allColSchema.Columns[i])
break
}
}
}
return schema
}
func getPhysicalID(t table.Table, isGlobalIndex bool) (physicalID int64, isPartition bool) {
tblInfo := t.Meta()
if !isGlobalIndex && tblInfo.GetPartitionInfo() != nil {
pid := t.(table.PhysicalTable).GetPhysicalID()
return pid, true
}
return tblInfo.ID, false
}
func tryGetPkExtraColumn(sv *variable.SessionVars, tblInfo *model.TableInfo) (*model.ColumnInfo, *expression.Column, bool) {
if tblInfo.IsCommonHandle || tblInfo.PKIsHandle {
return nil, nil, false
}
info := model.NewExtraHandleColInfo()
expCol := &expression.Column{
RetType: types.NewFieldType(mysql.TypeLonglong),
UniqueID: sv.AllocPlanColumnID(),
ID: model.ExtraHandleID,
}
return info, expCol, true
}
func tryGetCommonHandleCols(t table.Table, allColSchema *expression.Schema) ([]*table.Column, []*expression.Column, bool) {
tblInfo := t.Meta()
if !tblInfo.IsCommonHandle {
return nil, nil, false
}
pk := tables.FindPrimaryIndex(tblInfo)
commonHandleCols, _ := util.IndexInfo2FullCols(tblInfo.Columns, allColSchema.Columns, pk)
commonHandelColInfos := tables.TryGetCommonPkColumns(t)
return commonHandelColInfos, commonHandleCols, true
}
func tryGetPkHandleCol(tblInfo *model.TableInfo, allColSchema *expression.Schema) (*model.ColumnInfo, *expression.Column, bool) {
if !tblInfo.PKIsHandle {
return nil, nil, false
}
for i, c := range tblInfo.Columns {
if mysql.HasPriKeyFlag(c.GetFlag()) {
return c, allColSchema.Columns[i], true
}
}
return nil, nil, false
}
func (b *PlanBuilder) buildPhysicalIndexLookUpReaders(ctx context.Context, dbName ast.CIStr, tbl table.Table, indices []table.Index) ([]base.Plan, []*model.IndexInfo, error) {
tblInfo := tbl.Meta()
// get index information
indexInfos := make([]*model.IndexInfo, 0, len(tblInfo.Indices))
indexLookUpReaders := make([]base.Plan, 0, len(tblInfo.Indices))
check := b.isForUpdateRead || b.ctx.GetSessionVars().IsIsolation(ast.ReadCommitted)
check = check && b.ctx.GetSessionVars().ConnectionID > 0
var latestIndexes map[int64]*model.IndexInfo
var err error
for _, idx := range indices {
idxInfo := idx.Meta()
if tblInfo.IsCommonHandle && idxInfo.Primary {
// Skip checking clustered index.
continue
}
if idxInfo.State != model.StatePublic {
logutil.Logger(ctx).Info("build physical index lookup reader, the index isn't public",
zap.String("index", idxInfo.Name.O),
zap.Stringer("state", idxInfo.State),
zap.String("table", tblInfo.Name.O))
continue
}
if check && latestIndexes == nil {
latestIndexes, check, err = domainmisc.GetLatestIndexInfo(b.ctx, tblInfo.ID, b.is.SchemaMetaVersion())
if err != nil {
return nil, nil, err
}
}
if check {
if latestIndex, ok := latestIndexes[idxInfo.ID]; !ok || latestIndex.State != model.StatePublic {
forUpdateState := model.StateNone
if ok {
forUpdateState = latestIndex.State
}
logutil.Logger(ctx).Info("build physical index lookup reader, the index isn't public in forUpdateRead",
zap.String("index", idxInfo.Name.O),
zap.Stringer("state", idxInfo.State),
zap.Stringer("forUpdateRead state", forUpdateState),
zap.String("table", tblInfo.Name.O))
continue
}
}
indexInfos = append(indexInfos, idxInfo)
// For partition tables except global index.
if pi := tbl.Meta().GetPartitionInfo(); pi != nil && !idxInfo.Global {
for _, def := range pi.Definitions {
t := tbl.(table.PartitionedTable).GetPartition(def.ID)
reader, err := b.buildPhysicalIndexLookUpReader(ctx, dbName, t, idxInfo)
if err != nil {
return nil, nil, err
}
indexLookUpReaders = append(indexLookUpReaders, reader)
}
continue
}
// For non-partition tables.
reader, err := b.buildPhysicalIndexLookUpReader(ctx, dbName, tbl, idxInfo)
if err != nil {
return nil, nil, err
}
indexLookUpReaders = append(indexLookUpReaders, reader)
}
if len(indexLookUpReaders) == 0 {
return nil, nil, nil
}
return indexLookUpReaders, indexInfos, nil
}
func (b *PlanBuilder) buildAdminCheckTable(ctx context.Context, as *ast.AdminStmt) (*CheckTable, error) {
if len(as.Tables) > 1 {
return nil, errors.New("admin check only supports one table at a time")
}
tblName := as.Tables[0]
tnW := b.resolveCtx.GetTableName(tblName)
tableInfo := tnW.TableInfo
tbl, ok := b.is.TableByID(ctx, tableInfo.ID)
if !ok {
return nil, infoschema.ErrTableNotExists.FastGenByArgs(tnW.DBInfo.Name.O, tableInfo.Name.O)
}
p := &CheckTable{
DBName: tblName.Schema.O,
Table: tbl,
}
var readerPlans []base.Plan
var indexInfos []*model.IndexInfo
var err error
if as.Tp == ast.AdminCheckIndex {
// get index information
var idx table.Index
idxName := strings.ToLower(as.Index)
for _, index := range tbl.Indices() {
if index.Meta().Name.L == idxName {
idx = index
break
}
}
if idx == nil {
return nil, errors.Errorf("secondary index %s does not exist", as.Index)
}
if idx.Meta().State != model.StatePublic {
return nil, errors.Errorf("index %s state %s isn't public", as.Index, idx.Meta().State)
}
p.CheckIndex = true
readerPlans, indexInfos, err = b.buildPhysicalIndexLookUpReaders(ctx, tblName.Schema, tbl, []table.Index{idx})
} else {
readerPlans, indexInfos, err = b.buildPhysicalIndexLookUpReaders(ctx, tblName.Schema, tbl, tbl.Indices())
}
if err != nil {
return nil, errors.Trace(err)
}
readers := make([]*physicalop.PhysicalIndexLookUpReader, 0, len(readerPlans))
for _, plan := range readerPlans {
readers = append(readers, plan.(*physicalop.PhysicalIndexLookUpReader))
}
p.IndexInfos = indexInfos
p.IndexLookUpReaders = readers
return p, nil
}
func (b *PlanBuilder) buildCheckIndexSchema(tn *ast.TableName, indexName string) (*expression.Schema, types.NameSlice, error) {
schema := expression.NewSchema()
var names types.NameSlice
indexName = strings.ToLower(indexName)
tnW := b.resolveCtx.GetTableName(tn)
indicesInfo := tnW.TableInfo.Indices
cols := tnW.TableInfo.Cols()
for _, idxInfo := range indicesInfo {
if idxInfo.Name.L != indexName {
continue
}
for _, idxCol := range idxInfo.Columns {
col := cols[idxCol.Offset]
names = append(names, &types.FieldName{
ColName: idxCol.Name,
TblName: tn.Name,
DBName: tn.Schema,
})
schema.Append(&expression.Column{
RetType: &col.FieldType,
UniqueID: b.ctx.GetSessionVars().AllocPlanColumnID(),
ID: col.ID})
}
names = append(names, &types.FieldName{
ColName: ast.NewCIStr("extra_handle"),
TblName: tn.Name,
DBName: tn.Schema,
})
schema.Append(&expression.Column{
RetType: types.NewFieldType(mysql.TypeLonglong),
UniqueID: b.ctx.GetSessionVars().AllocPlanColumnID(),
ID: -1,
})
}
if schema.Len() == 0 {
return nil, nil, errors.Errorf("index %s not found", indexName)
}
return schema, names, nil
}
// getColsInfo returns the info of index columns, normal columns and primary key.
func (b *PlanBuilder) getColsInfo(tn *ast.TableName) (indicesInfo []*model.IndexInfo, colsInfo []*model.ColumnInfo) {
tnW := b.resolveCtx.GetTableName(tn)
tbl := tnW.TableInfo
for _, col := range tbl.Columns {
// The virtual column will not store any data in TiKV, so it should be ignored when collect statistics
if col.IsVirtualGenerated() {
continue
}
if mysql.HasPriKeyFlag(col.GetFlag()) && tbl.HasClusteredIndex() {
continue
}
colsInfo = append(colsInfo, col)
}
for _, idx := range tnW.TableInfo.Indices {
if idx.State == model.StatePublic {
indicesInfo = append(indicesInfo, idx)
}
}
return
}
// BuildHandleColsForAnalyze returns HandleCols for ANALYZE.
func BuildHandleColsForAnalyze(
_ base.PlanContext, tblInfo *model.TableInfo, allColumns bool, colsInfo []*model.ColumnInfo,
) util.HandleCols {
var handleCols util.HandleCols
switch {
case tblInfo.PKIsHandle:
pkCol := tblInfo.GetPkColInfo()
var index int
if allColumns {
// If all the columns need to be analyzed, we just set index to pkCol.Offset.
index = pkCol.Offset
} else {
// If only a part of the columns need to be analyzed, we need to set index according to colsInfo.
index = getColOffsetForAnalyze(colsInfo, pkCol.ID)
}
handleCols = util.NewIntHandleCols(&expression.Column{
ID: pkCol.ID,
RetType: &pkCol.FieldType,
Index: index,
})
case tblInfo.IsCommonHandle:
pkIdx := tables.FindPrimaryIndex(tblInfo)
pkColLen := len(pkIdx.Columns)
columns := make([]*expression.Column, pkColLen)
for i := range pkColLen {
colInfo := tblInfo.Columns[pkIdx.Columns[i].Offset]
var index int
if allColumns {
// If all the columns need to be analyzed, we just set index to colInfo.Offset.
index = colInfo.Offset
} else {
// If only a part of the columns need to be analyzed, we need to set index according to colsInfo.
index = getColOffsetForAnalyze(colsInfo, colInfo.ID)
}
columns[i] = &expression.Column{
ID: colInfo.ID,
RetType: &colInfo.FieldType,
Index: index,
}
}
// We don't modify IndexColumn.Offset for CommonHandleCols.idxInfo according to colsInfo. There are two reasons.
// The first reason is that we use Column.Index of CommonHandleCols.columns, rather than IndexColumn.Offset, to get
// column value from row samples when calling (*CommonHandleCols).BuildHandleByDatums in (*AnalyzeColumnsExec).buildSamplingStats.
// The second reason is that in (cb *CommonHandleCols).BuildHandleByDatums, tablecodec.TruncateIndexValues(cb.tblInfo, cb.idxInfo, datumBuf)
// is called, which asks that IndexColumn.Offset of cb.idxInfo must be according to cb,tblInfo.
// TODO: find a better way to find handle columns in ANALYZE rather than use Column.Index
handleCols = util.NewCommonHandlesColsWithoutColsAlign(tblInfo, pkIdx, columns)
}
return handleCols
}
// GetPhysicalIDsAndPartitionNames returns physical IDs and names of these partitions.
func GetPhysicalIDsAndPartitionNames(tblInfo *model.TableInfo, partitionNames []ast.CIStr) ([]int64, []string, error) {
pi := tblInfo.GetPartitionInfo()
if pi == nil {
if len(partitionNames) != 0 {
return nil, nil, errors.Trace(dbterror.ErrPartitionMgmtOnNonpartitioned)
}
return []int64{tblInfo.ID}, []string{""}, nil
}
// If the PartitionNames is empty, we will return all partitions.
if len(partitionNames) == 0 {
ids := make([]int64, 0, len(pi.Definitions))
names := make([]string, 0, len(pi.Definitions))
for _, def := range pi.Definitions {
ids = append(ids, def.ID)
names = append(names, def.Name.O)
}
return ids, names, nil
}
ids := make([]int64, 0, len(partitionNames))
names := make([]string, 0, len(partitionNames))
for _, name := range partitionNames {
found := false
for _, def := range pi.Definitions {
if def.Name.L == name.L {
found = true
ids = append(ids, def.ID)
names = append(names, def.Name.O)
break
}
}
if !found {
return nil, nil, fmt.Errorf("can not found the specified partition name %s in the table definition", name.O)
}
}
return ids, names, nil
}
type calcOnceMap struct {
data map[int64]struct{}
calculated bool
}
// addColumnsWithVirtualExprs adds columns to cols.data and processes their virtual expressions recursively.
// columnSelector is a function that determines which columns to include and their virtual expressions.
func (b *PlanBuilder) addColumnsWithVirtualExprs(tbl *resolve.TableNameW, cols *calcOnceMap, columnSelector func([]*expression.Column) []expression.Expression) error {
intest.Assert(cols.data != nil, "cols.data should not be nil before adding columns")
tblInfo := tbl.TableInfo
columns, _, err := expression.ColumnInfos2ColumnsAndNames(b.ctx.GetExprCtx(), tbl.Schema, tbl.Name, tblInfo.Columns, tblInfo)
if err != nil {
return err
}
virtualExprs := columnSelector(columns)
relatedCols := expression.GetUniqueIDToColumnMap()
defer expression.PutUniqueIDToColumnMap(relatedCols)
for len(virtualExprs) > 0 {
expression.ExtractColumnsMapFromExpressionsWithReusedMap(relatedCols, nil, virtualExprs...)
virtualExprs = virtualExprs[:0]
for _, col := range relatedCols {
cols.data[col.ID] = struct{}{}
if col.VirtualExpr != nil {
virtualExprs = append(virtualExprs, col.VirtualExpr)
}
}
clear(relatedCols)
}
return nil
}
// getMustAnalyzedColumns puts the columns whose statistics must be collected into `cols` if `cols` has not been calculated.
func (b *PlanBuilder) getMustAnalyzedColumns(tbl *resolve.TableNameW, cols *calcOnceMap) (map[int64]struct{}, error) {
if cols.calculated {
return cols.data, nil
}
tblInfo := tbl.TableInfo
cols.data = make(map[int64]struct{}, len(tblInfo.Columns))
if len(tblInfo.Indices) > 0 {
// Add indexed columns.
// Some indexed columns are generated columns so we also need to add the columns that make up those generated columns.
err := b.addColumnsWithVirtualExprs(tbl, cols, func(columns []*expression.Column) []expression.Expression {
virtualExprs := make([]expression.Expression, 0, len(tblInfo.Columns))
for _, idx := range tblInfo.Indices {
// for an index in state public, we can always analyze it for internal analyze session.
// for an index in state WriteReorg, we can only analyze it under variable EnableDDLAnalyze is on.
indexStateAnalyzable := idx.State == model.StatePublic ||
(idx.State == model.StateWriteReorganization && b.ctx.GetSessionVars().EnableDDLAnalyzeExecOpt)
// for mv index and ci index fail it first, then analyze those analyzable indexes.
if idx.MVIndex || idx.IsColumnarIndex() || !indexStateAnalyzable {
continue
}
for _, idxCol := range idx.Columns {
colInfo := tblInfo.Columns[idxCol.Offset]
cols.data[colInfo.ID] = struct{}{}
if expr := columns[idxCol.Offset].VirtualExpr; expr != nil {
virtualExprs = append(virtualExprs, expr)
}
}
}
return virtualExprs
})
if err != nil {
return nil, err
}
}
if tblInfo.PKIsHandle {
pkCol := tblInfo.GetPkColInfo()
cols.data[pkCol.ID] = struct{}{}
}
if b.ctx.GetSessionVars().EnableExtendedStats {
// Add the columns related to extended stats.
// TODO: column_ids read from mysql.stats_extended in optimization phase may be different from that in execution phase((*Handle).BuildExtendedStats)
// if someone inserts data into mysql.stats_extended between the two time points, the new added extended stats may not be computed.
statsHandle := domain.GetDomain(b.ctx).StatsHandle()
extendedStatsColIDs, err := statsHandle.CollectColumnsInExtendedStats(tblInfo.ID)
if err != nil {
return nil, err
}
for _, colID := range extendedStatsColIDs {
cols.data[colID] = struct{}{}
}
}
cols.calculated = true
return cols.data, nil
}
// getPredicateColumns gets the columns used in predicates.
func (b *PlanBuilder) getPredicateColumns(tbl *resolve.TableNameW, cols *calcOnceMap) (map[int64]struct{}, error) {
// Already calculated in the previous call.
if cols.calculated {
return cols.data, nil
}
tblInfo := tbl.TableInfo
cols.data = make(map[int64]struct{}, len(tblInfo.Columns))
do := domain.GetDomain(b.ctx)
h := do.StatsHandle()
colList, err := h.GetPredicateColumns(tblInfo.ID)
if err != nil {
return nil, err
}
if len(colList) == 0 {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(
errors.NewNoStackErrorf(
"No predicate column has been collected yet for table %s.%s, so only indexes and the columns composing the indexes will be analyzed",
tbl.Schema.L,
tbl.Name.L,
),
)
} else {
// Some predicate columns are generated columns so we also need to add the columns that make up those generated columns.
err := b.addColumnsWithVirtualExprs(tbl, cols, func(columns []*expression.Column) []expression.Expression {
virtualExprs := make([]expression.Expression, 0, len(tblInfo.Columns))
for _, id := range colList {
columnInfo := tblInfo.GetColumnByID(id)
intest.Assert(columnInfo != nil, "column %d not found in table %s.%s", id, tbl.Schema.L, tbl.Name.L)
if columnInfo == nil {
statslogutil.StatsLogger().Warn("Column not found in table while getting predicate columns",
zap.Int64("columnID", id),
zap.String("tableSchema", tbl.Schema.L),
zap.String("table", tblInfo.Name.O),
)
// This should not happen, but we handle it gracefully.
continue
}
cols.data[columnInfo.ID] = struct{}{}
if expr := columns[columnInfo.Offset].VirtualExpr; expr != nil {
virtualExprs = append(virtualExprs, expr)
}
}
return virtualExprs
})
if err != nil {
return nil, err
}
}
cols.calculated = true
return cols.data, nil
}
func getAnalyzeColumnList(specifiedColumns []ast.CIStr, tbl *resolve.TableNameW) ([]*model.ColumnInfo, error) {
colList := make([]*model.ColumnInfo, 0, len(specifiedColumns))
for _, colName := range specifiedColumns {
colInfo := model.FindColumnInfo(tbl.TableInfo.Columns, colName.L)
if colInfo == nil {
return nil, plannererrors.ErrAnalyzeMissColumn.GenWithStackByArgs(colName.O, tbl.TableInfo.Name.O)
}
colList = append(colList, colInfo)
}
return colList, nil
}
// getFullAnalyzeColumnsInfo decides which columns need to be analyzed.
// The first return value is the columns which need to be analyzed and the second return value is the columns which need to
// be record in mysql.analyze_options(only for the case of analyze table t columns c1, .., cn).
func (b *PlanBuilder) getFullAnalyzeColumnsInfo(
tbl *resolve.TableNameW,
columnChoice ast.ColumnChoice,
specifiedCols []*model.ColumnInfo,
predicateCols, mustAnalyzedCols *calcOnceMap,
mustAllColumns bool,
warning bool,
) (_, _ []*model.ColumnInfo, _ error) {
if mustAllColumns && warning && (columnChoice == ast.PredicateColumns || columnChoice == ast.ColumnList) {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("Table %s.%s has version 1 statistics so all the columns must be analyzed to overwrite the current statistics", tbl.Schema.L, tbl.Name.L))
}
switch columnChoice {
case ast.DefaultChoice:
columnOptions := vardef.AnalyzeColumnOptions.Load()
switch columnOptions {
case ast.AllColumns.String():
return tbl.TableInfo.Columns, nil, nil
case ast.PredicateColumns.String():
columns, err := b.getColumnsBasedOnPredicateColumns(
tbl,
predicateCols,
mustAnalyzedCols,
mustAllColumns,
)
if err != nil {
return nil, nil, err
}
return columns, nil, nil
default:
// Usually, this won't happen.
statslogutil.StatsLogger().Warn("Unknown default column choice, analyze all columns", zap.String("choice", columnOptions))
return tbl.TableInfo.Columns, nil, nil
}
case ast.AllColumns:
return tbl.TableInfo.Columns, nil, nil
case ast.PredicateColumns:
columns, err := b.getColumnsBasedOnPredicateColumns(
tbl,
predicateCols,
mustAnalyzedCols,
mustAllColumns,
)
if err != nil {
return nil, nil, err
}
return columns, nil, nil
case ast.ColumnList:
colSet := getColumnSetFromSpecifiedCols(specifiedCols)
mustAnalyzed, err := b.getMustAnalyzedColumns(tbl, mustAnalyzedCols)
if err != nil {
return nil, nil, err
}
if warning {
missing := getMissingColumns(colSet, mustAnalyzed)
if len(missing) > 0 {
missingNames := getColumnNamesFromIDs(tbl.TableInfo.Columns, missing)
warningMsg := fmt.Sprintf("Columns %s are missing in ANALYZE but their stats are needed for calculating stats for indexes/primary key/extended stats", strings.Join(missingNames, ","))
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError(warningMsg))
}
}
colSet = combineColumnSets(colSet, mustAnalyzed)
colList := getColumnListFromSet(tbl.TableInfo.Columns, colSet)
if mustAllColumns {
return tbl.TableInfo.Columns, colList, nil
}
return colList, colList, nil
}
return nil, nil, nil
}
func (b *PlanBuilder) getColumnsBasedOnPredicateColumns(
tbl *resolve.TableNameW,
predicateCols, mustAnalyzedCols *calcOnceMap,
rewriteAllStatsNeeded bool,
) ([]*model.ColumnInfo, error) {
if rewriteAllStatsNeeded {
return tbl.TableInfo.Columns, nil
}
predicate, err := b.getPredicateColumns(tbl, predicateCols)
if err != nil {
return nil, err
}
mustAnalyzed, err := b.getMustAnalyzedColumns(tbl, mustAnalyzedCols)
if err != nil {
return nil, err
}
colSet := combineColumnSets(predicate, mustAnalyzed)
return getColumnListFromSet(tbl.TableInfo.Columns, colSet), nil
}
// Helper function to combine two column sets.
func combineColumnSets(sets ...map[int64]struct{}) map[int64]struct{} {
result := make(map[int64]struct{})
for _, set := range sets {
for colID := range set {
result[colID] = struct{}{}
}
}
return result
}
// Helper function to extract column IDs from specified columns.
func getColumnSetFromSpecifiedCols(cols []*model.ColumnInfo) map[int64]struct{} {
colSet := make(map[int64]struct{}, len(cols))
for _, colInfo := range cols {
colSet[colInfo.ID] = struct{}{}
}
return colSet
}
// Helper function to get missing columns from a set.
func getMissingColumns(colSet, mustAnalyzed map[int64]struct{}) map[int64]struct{} {
missing := make(map[int64]struct{})
for colID := range mustAnalyzed {
if _, ok := colSet[colID]; !ok {
missing[colID] = struct{}{}
}
}
return missing
}
// Helper function to get column names from IDs.
func getColumnNamesFromIDs(columns []*model.ColumnInfo, colIDs map[int64]struct{}) []string {
var missingNames []string
for _, col := range columns {
if _, ok := colIDs[col.ID]; ok {
missingNames = append(missingNames, col.Name.O)
}
}
return missingNames
}
// Helper function to get a list of column infos from a set of column IDs.
func getColumnListFromSet(columns []*model.ColumnInfo, colSet map[int64]struct{}) []*model.ColumnInfo {
colList := make([]*model.ColumnInfo, 0, len(colSet))
for _, colInfo := range columns {
if _, ok := colSet[colInfo.ID]; ok {
colList = append(colList, colInfo)
}
}
return colList
}
func getColOffsetForAnalyze(colsInfo []*model.ColumnInfo, colID int64) int {
for i, col := range colsInfo {
if colID == col.ID {
return i
}
}
return -1
}
// getModifiedIndexesInfoForAnalyze returns indexesInfo for ANALYZE.
// 1. If allColumns is true, we just return public indexes in tblInfo.Indices.
// 2. If allColumns is false, colsInfo indicate the columns whose stats need to be collected. colsInfo is a subset of tbl.Columns. For each public index
// in tblInfo.Indices, index.Columns[i].Offset is set according to tblInfo.Columns. Since we decode row samples according to colsInfo rather than tbl.Columns
// in the execution phase of ANALYZE, we need to modify index.Columns[i].Offset according to colInfos.
// TODO: find a better way to find indexed columns in ANALYZE rather than use IndexColumn.Offset
// For multi-valued index, we need to collect it separately here and analyze it as independent index analyze task.
// For a special global index, we also need to analyze it as independent index analyze task.
// See comments for AnalyzeResults.ForMVIndex for more details.
func getModifiedIndexesInfoForAnalyze(
sCtx base.PlanContext,
tblInfo *model.TableInfo,
allColumns bool,
colsInfo []*model.ColumnInfo,
) (idxsInfo, independentIdxsInfo, specialGlobalIdxsInfo []*model.IndexInfo) {
idxsInfo = make([]*model.IndexInfo, 0, len(tblInfo.Indices))
independentIdxsInfo = make([]*model.IndexInfo, 0)
specialGlobalIdxsInfo = make([]*model.IndexInfo, 0)
for _, originIdx := range tblInfo.Indices {
// for an index in state public, we can always analyze it for internal analyze session.
// for an index in state WriteReorg, we can only analyze it under variable EnableDDLAnalyze is on.
indexStateAnalyzable := originIdx.State == model.StatePublic ||
(originIdx.State == model.StateWriteReorganization && sCtx.GetSessionVars().EnableDDLAnalyzeExecOpt)
if !indexStateAnalyzable {
continue
}
if handleutil.IsSpecialGlobalIndex(originIdx, tblInfo) {
specialGlobalIdxsInfo = append(specialGlobalIdxsInfo, originIdx)
continue
}
if originIdx.MVIndex {
independentIdxsInfo = append(independentIdxsInfo, originIdx)
continue
}
if originIdx.IsColumnarIndex() {
sCtx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing columnar index is not supported, skip %s", originIdx.Name.L))
continue
}
if allColumns {
// If all the columns need to be analyzed, we don't need to modify IndexColumn.Offset.
idxsInfo = append(idxsInfo, originIdx)
continue
}
// If only a part of the columns need to be analyzed, we need to set IndexColumn.Offset according to colsInfo.
idx := originIdx.Clone()
for i, idxCol := range idx.Columns {
colID := tblInfo.Columns[idxCol.Offset].ID
idx.Columns[i].Offset = getColOffsetForAnalyze(colsInfo, colID)
}
idxsInfo = append(idxsInfo, idx)
}
return idxsInfo, independentIdxsInfo, specialGlobalIdxsInfo
}
// filterSkipColumnTypes filters out columns whose types are in the skipTypes list.
func (b *PlanBuilder) filterSkipColumnTypes(origin []*model.ColumnInfo, tbl *resolve.TableNameW, mustAnalyzedCols *calcOnceMap) (result []*model.ColumnInfo, skipCol []*model.ColumnInfo) {
// For auto-analyze, it uses @@global.tidb_analyze_skip_column_types to obtain the skipTypes list.
// This is already handled before executing the query by the CallWithSCtx utility function.
skipTypes := b.ctx.GetSessionVars().AnalyzeSkipColumnTypes
mustAnalyze, err1 := b.getMustAnalyzedColumns(tbl, mustAnalyzedCols)
if err1 != nil {
logutil.BgLogger().Error("getting must-analyzed columns failed", zap.Error(err1))
result = origin
return
}
// If one column's type is in the skipTypes list and it doesn't exist in mustAnalyzedCols, we will skip it.
for _, colInfo := range origin {
// Vector type is skip by hardcoded. Just because that collecting it is meanless for current TiDB.
if colInfo.FieldType.GetType() == mysql.TypeTiDBVectorFloat32 {
skipCol = append(skipCol, colInfo)
continue
}
_, skip := skipTypes[types.TypeToStr(colInfo.FieldType.GetType(), colInfo.FieldType.GetCharset())]
// Currently, if the column exists in some index(except MV Index), we need to bring the column's sample values
// into TiDB to build the index statistics.
_, keep := mustAnalyze[colInfo.ID]
if skip && !keep {
skipCol = append(skipCol, colInfo)
continue
}
result = append(result, colInfo)
}
skipColNameMap := make(map[string]struct{}, len(skipCol))
for _, colInfo := range skipCol {
skipColNameMap[colInfo.Name.L] = struct{}{}
}
// Filter out generated columns that depend on skipped columns
filteredResult := make([]*model.ColumnInfo, 0, len(result))
for _, colInfo := range result {
shouldSkip := false
if colInfo.IsGenerated() {
// Check if any dependency is in the skip list
for depName := range colInfo.Dependences {
if _, exists := skipColNameMap[depName]; exists {
skipCol = append(skipCol, colInfo)
shouldSkip = true
break
}
}
}
if !shouldSkip {
filteredResult = append(filteredResult, colInfo)
}
}
result = filteredResult
return
}
// This function is to check whether all indexes is special global index or not.
// A special global index is an index that is both a global index and an expression index or a prefix index.
func checkIsAllSpecialGlobalIndex(as *ast.AnalyzeTableStmt, tbl *resolve.TableNameW) (bool, error) {
isAnalyzeTable := len(as.PartitionNames) == 0
// For `Analyze table t index`
if as.IndexFlag && len(as.IndexNames) == 0 {
for _, idx := range tbl.TableInfo.Indices {
if idx.State != model.StatePublic {
continue
}
if !handleutil.IsSpecialGlobalIndex(idx, tbl.TableInfo) {
return false, nil
}
// For `Analyze table t partition p0 index`
if !isAnalyzeTable {
return false, errors.NewNoStackErrorf("Analyze global index '%s' can't work with analyze specified partitions", idx.Name.O)
}
}
} else {
for _, idxName := range as.IndexNames {
idx := tbl.TableInfo.FindIndexByName(idxName.L)
if idx == nil || idx.State != model.StatePublic {
return false, plannererrors.ErrAnalyzeMissIndex.GenWithStackByArgs(idxName.O, tbl.Name.O)
}
if !handleutil.IsSpecialGlobalIndex(idx, tbl.TableInfo) {
return false, nil
}
// For `Analyze table t partition p0 index idx0`
if !isAnalyzeTable {
return false, errors.NewNoStackErrorf("Analyze global index '%s' can't work with analyze specified partitions", idx.Name.O)
}
}
}
return true, nil
}
func (b *PlanBuilder) buildAnalyzeFullSamplingTask(
as *ast.AnalyzeTableStmt,
analyzePlan *Analyze,
physicalIDs []int64,
partitionNames []string,
tbl *resolve.TableNameW,
version int,
persistOpts bool,
) error {
// Version 2 doesn't support incremental analyze.
// And incremental analyze will be deprecated in the future.
if as.Incremental {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError("The version 2 stats would ignore the INCREMENTAL keyword and do full sampling"))
}
isAnalyzeTable := len(as.PartitionNames) == 0
allSpecialGlobalIndex, err := checkIsAllSpecialGlobalIndex(as, tbl)
if err != nil {
return err
}
astOpts, err := handleAnalyzeOptionsV2(as.AnalyzeOpts)
if err != nil {
return err
}
// Get all column info which need to be analyzed.
astColList, err := getAnalyzeColumnList(as.ColumnNames, tbl)
if err != nil {
return err
}
var predicateCols, mustAnalyzedCols calcOnceMap
ver := version
statsHandle := domain.GetDomain(b.ctx).StatsHandle()
// If the statistics of the table is version 1, we must analyze all columns to overwrites all of old statistics.
mustAllColumns := !statsHandle.CheckAnalyzeVersion(tbl.TableInfo, physicalIDs, &ver)
astColsInfo, _, err := b.getFullAnalyzeColumnsInfo(tbl, as.ColumnChoice, astColList, &predicateCols, &mustAnalyzedCols, mustAllColumns, true)
if err != nil {
return err
}
optionsMap, colsInfoMap, err := b.genV2AnalyzeOptions(persistOpts, tbl, isAnalyzeTable, physicalIDs, astOpts, as.ColumnChoice, astColList, &predicateCols, &mustAnalyzedCols, mustAllColumns)
if err != nil {
return err
}
maps.Copy(analyzePlan.OptionsMap, optionsMap)
var indexes, independentIndexes, specialGlobalIndexes []*model.IndexInfo
needAnalyzeCols := !(as.IndexFlag && allSpecialGlobalIndex)
if needAnalyzeCols {
if as.IndexFlag {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("The version 2 would collect all statistics not only the selected indexes"))
}
// Build tasks for each partition.
for i, id := range physicalIDs {
physicalID := id
if id == tbl.TableInfo.ID {
id = statistics.NonPartitionTableID
}
info := AnalyzeInfo{
DBName: tbl.Schema.O,
TableName: tbl.Name.O,
PartitionName: partitionNames[i],
TableID: statistics.AnalyzeTableID{TableID: tbl.TableInfo.ID, PartitionID: id},
StatsVersion: version,
}
if optsV2, ok := optionsMap[physicalID]; ok {
info.V2Options = &optsV2
}
execColsInfo := astColsInfo
if colsInfo, ok := colsInfoMap[physicalID]; ok {
execColsInfo = colsInfo
}
var skipColsInfo []*model.ColumnInfo
execColsInfo, skipColsInfo = b.filterSkipColumnTypes(execColsInfo, tbl, &mustAnalyzedCols)
allColumns := len(tbl.TableInfo.Columns) == len(execColsInfo)
indexes, independentIndexes, specialGlobalIndexes = getModifiedIndexesInfoForAnalyze(b.ctx, tbl.TableInfo, allColumns, execColsInfo)
handleCols := BuildHandleColsForAnalyze(b.ctx, tbl.TableInfo, allColumns, execColsInfo)
newTask := AnalyzeColumnsTask{
HandleCols: handleCols,
ColsInfo: execColsInfo,
AnalyzeInfo: info,
TblInfo: tbl.TableInfo,
Indexes: indexes,
SkipColsInfo: skipColsInfo,
}
if newTask.HandleCols == nil {
extraCol := model.NewExtraHandleColInfo()
// Always place _tidb_rowid at the end of colsInfo, this is corresponding to logics in `analyzeColumnsPushdown`.
newTask.ColsInfo = append(newTask.ColsInfo, extraCol)
newTask.HandleCols = util.NewIntHandleCols(colInfoToColumn(extraCol, len(newTask.ColsInfo)-1))
}
analyzePlan.ColTasks = append(analyzePlan.ColTasks, newTask)
for _, indexInfo := range independentIndexes {
newIdxTask := AnalyzeIndexTask{
IndexInfo: indexInfo,
TblInfo: tbl.TableInfo,
AnalyzeInfo: info,
}
analyzePlan.IdxTasks = append(analyzePlan.IdxTasks, newIdxTask)
}
}
}
if isAnalyzeTable {
if needAnalyzeCols {
// When `needAnalyzeCols == true`, non-global indexes already covered by previous loop,
// deal with global index here.
for _, indexInfo := range specialGlobalIndexes {
analyzePlan.IdxTasks = append(analyzePlan.IdxTasks, generateIndexTasks(indexInfo, as, tbl.TableInfo, nil, nil, version)...)
}
} else {
// For `analyze table t index idx1[, idx2]` and all indexes are global index.
for _, idxName := range as.IndexNames {
idx := tbl.TableInfo.FindIndexByName(idxName.L)
if idx == nil || !handleutil.IsSpecialGlobalIndex(idx, tbl.TableInfo) {
continue
}
analyzePlan.IdxTasks = append(analyzePlan.IdxTasks, generateIndexTasks(idx, as, tbl.TableInfo, nil, nil, version)...)
}
}
}
return nil
}
func (b *PlanBuilder) genV2AnalyzeOptions(
persist bool,
tbl *resolve.TableNameW,
isAnalyzeTable bool,
physicalIDs []int64,
astOpts map[ast.AnalyzeOptionType]uint64,
astColChoice ast.ColumnChoice,
astColList []*model.ColumnInfo,
predicateCols, mustAnalyzedCols *calcOnceMap,
mustAllColumns bool,
) (map[int64]V2AnalyzeOptions, map[int64][]*model.ColumnInfo, error) {
optionsMap := make(map[int64]V2AnalyzeOptions, len(physicalIDs))
colsInfoMap := make(map[int64][]*model.ColumnInfo, len(physicalIDs))
if !persist {
return optionsMap, colsInfoMap, nil
}
// In dynamic mode, we collect statistics for all partitions of the table as a global statistics.
// In static mode, each partition generates its own execution plan, which is then combined with PartitionUnion.
// Because the plan is generated for each partition individually, each partition uses its own statistics;
// In dynamic mode, there is no partitioning, and a global plan is generated for the whole table, so a global statistic is needed;
dynamicPrune := variable.PartitionPruneMode(b.ctx.GetSessionVars().PartitionPruneMode.Load()) == variable.Dynamic
if !isAnalyzeTable && dynamicPrune && (len(astOpts) > 0 || astColChoice != ast.DefaultChoice) {
astOpts = make(map[ast.AnalyzeOptionType]uint64, 0)
astColChoice = ast.DefaultChoice
astColList = make([]*model.ColumnInfo, 0)
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError("Ignore columns and options when analyze partition in dynamic mode"))
}
// Get the analyze options which are saved in mysql.analyze_options.
tblSavedOpts, tblSavedColChoice, tblSavedColList, err := b.getSavedAnalyzeOpts(tbl.TableInfo.ID, tbl.TableInfo)
if err != nil {
return nil, nil, err
}
tblOpts := tblSavedOpts
tblColChoice := tblSavedColChoice
tblColList := tblSavedColList
if isAnalyzeTable {
tblOpts = mergeAnalyzeOptions(astOpts, tblSavedOpts)
tblColChoice, tblColList = pickColumnList(astColChoice, astColList, tblSavedColChoice, tblSavedColList)
}
tblFilledOpts := fillAnalyzeOptionsV2(tblOpts)
tblColsInfo, tblColList, err := b.getFullAnalyzeColumnsInfo(tbl, tblColChoice, tblColList, predicateCols, mustAnalyzedCols, mustAllColumns, false)
if err != nil {
return nil, nil, err
}
tblAnalyzeOptions := V2AnalyzeOptions{
PhyTableID: tbl.TableInfo.ID,
RawOpts: tblOpts,
FilledOpts: tblFilledOpts,
ColChoice: tblColChoice,
ColumnList: tblColList,
IsPartition: false,
}
optionsMap[tbl.TableInfo.ID] = tblAnalyzeOptions
colsInfoMap[tbl.TableInfo.ID] = tblColsInfo
for _, physicalID := range physicalIDs {
// This is a partitioned table, we need to collect statistics for each partition.
if physicalID != tbl.TableInfo.ID {
// In dynamic mode, we collect statistics for all partitions of the table as a global statistics.
// So we use the same options as the table level.
if dynamicPrune {
parV2Options := V2AnalyzeOptions{
PhyTableID: physicalID,
RawOpts: tblOpts,
FilledOpts: tblFilledOpts,
ColChoice: tblColChoice,
ColumnList: tblColList,
IsPartition: true,
}
optionsMap[physicalID] = parV2Options
colsInfoMap[physicalID] = tblColsInfo
continue
}
parSavedOpts, parSavedColChoice, parSavedColList, err := b.getSavedAnalyzeOpts(physicalID, tbl.TableInfo)
if err != nil {
return nil, nil, err
}
// merge partition level options with table level options firstly
savedOpts := mergeAnalyzeOptions(parSavedOpts, tblSavedOpts)
savedColChoice, savedColList := pickColumnList(parSavedColChoice, parSavedColList, tblSavedColChoice, tblSavedColList)
// then merge statement level options
mergedOpts := mergeAnalyzeOptions(astOpts, savedOpts)
filledMergedOpts := fillAnalyzeOptionsV2(mergedOpts)
finalColChoice, mergedColList := pickColumnList(astColChoice, astColList, savedColChoice, savedColList)
finalColsInfo, finalColList, err := b.getFullAnalyzeColumnsInfo(tbl, finalColChoice, mergedColList, predicateCols, mustAnalyzedCols, mustAllColumns, false)
if err != nil {
return nil, nil, err
}
parV2Options := V2AnalyzeOptions{
PhyTableID: physicalID,
RawOpts: mergedOpts,
FilledOpts: filledMergedOpts,
ColChoice: finalColChoice,
ColumnList: finalColList,
}
optionsMap[physicalID] = parV2Options
colsInfoMap[physicalID] = finalColsInfo
}
}
return optionsMap, colsInfoMap, nil
}
// getSavedAnalyzeOpts gets the analyze options which are saved in mysql.analyze_options.
func (b *PlanBuilder) getSavedAnalyzeOpts(physicalID int64, tblInfo *model.TableInfo) (map[ast.AnalyzeOptionType]uint64, ast.ColumnChoice, []*model.ColumnInfo, error) {
analyzeOptions := map[ast.AnalyzeOptionType]uint64{}
exec := b.ctx.GetRestrictedSQLExecutor()
ctx := kv.WithInternalSourceType(context.Background(), kv.InternalTxnStats)
rows, _, err := exec.ExecRestrictedSQL(ctx, nil, "select sample_num,sample_rate,buckets,topn,column_choice,column_ids from mysql.analyze_options where table_id = %?", physicalID)
if err != nil {
return nil, ast.DefaultChoice, nil, err
}
if len(rows) <= 0 {
return analyzeOptions, ast.DefaultChoice, nil, nil
}
row := rows[0]
sampleNum := row.GetInt64(0)
if sampleNum > 0 {
analyzeOptions[ast.AnalyzeOptNumSamples] = uint64(sampleNum)
}
sampleRate := row.GetFloat64(1)
if sampleRate > 0 {
analyzeOptions[ast.AnalyzeOptSampleRate] = math.Float64bits(sampleRate)
}
buckets := row.GetInt64(2)
if buckets > 0 {
analyzeOptions[ast.AnalyzeOptNumBuckets] = uint64(buckets)
}
topn := row.GetInt64(3)
if topn >= 0 {
analyzeOptions[ast.AnalyzeOptNumTopN] = uint64(topn)
}
colType := row.GetEnum(4)
switch colType.Name {
case "ALL":
return analyzeOptions, ast.AllColumns, tblInfo.Columns, nil
case "LIST":
colIDStrs := strings.Split(row.GetString(5), ",")
colList := make([]*model.ColumnInfo, 0, len(colIDStrs))
for _, colIDStr := range colIDStrs {
colID, _ := strconv.ParseInt(colIDStr, 10, 64)
colInfo := model.FindColumnInfoByID(tblInfo.Columns, colID)
if colInfo != nil {
colList = append(colList, colInfo)
}
}
return analyzeOptions, ast.ColumnList, colList, nil
case "PREDICATE":
return analyzeOptions, ast.PredicateColumns, nil, nil
default:
return analyzeOptions, ast.DefaultChoice, nil, nil
}
}
func mergeAnalyzeOptions(stmtOpts map[ast.AnalyzeOptionType]uint64, savedOpts map[ast.AnalyzeOptionType]uint64) map[ast.AnalyzeOptionType]uint64 {
merged := map[ast.AnalyzeOptionType]uint64{}
for optType := range ast.AnalyzeOptionString {
if stmtOpt, ok := stmtOpts[optType]; ok {
merged[optType] = stmtOpt
} else if savedOpt, ok := savedOpts[optType]; ok {
merged[optType] = savedOpt
}
}
return merged
}
// pickColumnList picks the column list to be analyzed.
// If the column list is specified in the statement, we will use it.
func pickColumnList(astColChoice ast.ColumnChoice, astColList []*model.ColumnInfo, tblSavedColChoice ast.ColumnChoice, tblSavedColList []*model.ColumnInfo) (ast.ColumnChoice, []*model.ColumnInfo) {
if astColChoice != ast.DefaultChoice {
return astColChoice, astColList
}
return tblSavedColChoice, tblSavedColList
}
// buildAnalyzeTable constructs analyze tasks for each table.
func (b *PlanBuilder) buildAnalyzeTable(as *ast.AnalyzeTableStmt, opts map[ast.AnalyzeOptionType]uint64, version int) (base.Plan, error) {
p := &Analyze{Opts: opts}
p.OptionsMap = make(map[int64]V2AnalyzeOptions)
usePersistedOptions := vardef.PersistAnalyzeOptions.Load()
// Construct tasks for each table.
for _, tbl := range as.TableNames {
tnW := b.resolveCtx.GetTableName(tbl)
if tnW.TableInfo.IsView() {
return nil, errors.Errorf("analyze view %s is not supported now", tbl.Name.O)
}
if tnW.TableInfo.IsSequence() {
return nil, errors.Errorf("analyze sequence %s is not supported now", tbl.Name.O)
}
idxInfo, colInfo := b.getColsInfo(tbl)
physicalIDs, partitionNames, err := GetPhysicalIDsAndPartitionNames(tnW.TableInfo, as.PartitionNames)
if err != nil {
return nil, err
}
var commonHandleInfo *model.IndexInfo
if version == statistics.Version2 {
err = b.buildAnalyzeFullSamplingTask(as, p, physicalIDs, partitionNames, tnW, version, usePersistedOptions)
if err != nil {
return nil, err
}
continue
}
// Version 1 analyze.
if as.ColumnChoice == ast.PredicateColumns {
return nil, errors.Errorf("Only the version 2 of analyze supports analyzing predicate columns")
}
if as.ColumnChoice == ast.ColumnList {
return nil, errors.Errorf("Only the version 2 of analyze supports analyzing the specified columns")
}
for _, idx := range idxInfo {
// For prefix common handle. We don't use analyze mixed to handle it with columns. Because the full value
// is read by coprocessor, the prefix index would get wrong stats in this case.
if idx.Primary && tnW.TableInfo.IsCommonHandle && !idx.HasPrefixIndex() {
commonHandleInfo = idx
continue
}
if idx.MVIndex {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing multi-valued indexes is not supported, skip %s", idx.Name.L))
continue
}
if idx.IsColumnarIndex() {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing columnar index is not supported, skip %s", idx.Name.L))
continue
}
p.IdxTasks = append(p.IdxTasks, generateIndexTasks(idx, as, tnW.TableInfo, partitionNames, physicalIDs, version)...)
}
handleCols := BuildHandleColsForAnalyze(b.ctx, tnW.TableInfo, true, nil)
if len(colInfo) > 0 || handleCols != nil {
for i, id := range physicalIDs {
if id == tnW.TableInfo.ID {
id = statistics.NonPartitionTableID
}
info := AnalyzeInfo{
DBName: tbl.Schema.O,
TableName: tbl.Name.O,
PartitionName: partitionNames[i],
TableID: statistics.AnalyzeTableID{TableID: tnW.TableInfo.ID, PartitionID: id},
StatsVersion: version,
}
p.ColTasks = append(p.ColTasks, AnalyzeColumnsTask{
HandleCols: handleCols,
CommonHandleInfo: commonHandleInfo,
ColsInfo: colInfo,
AnalyzeInfo: info,
TblInfo: tnW.TableInfo,
})
}
}
}
return p, nil
}
func (b *PlanBuilder) buildAnalyzeIndex(as *ast.AnalyzeTableStmt, opts map[ast.AnalyzeOptionType]uint64, version int) (base.Plan, error) {
p := &Analyze{Opts: opts}
statsHandle := domain.GetDomain(b.ctx).StatsHandle()
if statsHandle == nil {
return nil, errors.Errorf("statistics hasn't been initialized, please try again later")
}
tnW := b.resolveCtx.GetTableName(as.TableNames[0])
tblInfo := tnW.TableInfo
physicalIDs, names, err := GetPhysicalIDsAndPartitionNames(tblInfo, as.PartitionNames)
if err != nil {
return nil, err
}
versionIsSame := statsHandle.CheckAnalyzeVersion(tblInfo, physicalIDs, &version)
if !versionIsSame {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackError("The analyze version from the session is not compatible with the existing statistics of the table. Use the existing version instead"))
}
if version == statistics.Version2 {
return b.buildAnalyzeTable(as, opts, version)
}
for _, idxName := range as.IndexNames {
if isPrimaryIndex(idxName) {
handleCols := BuildHandleColsForAnalyze(b.ctx, tblInfo, true, nil)
// FIXME: How about non-int primary key?
if handleCols != nil && handleCols.IsInt() {
for i, id := range physicalIDs {
if id == tblInfo.ID {
id = statistics.NonPartitionTableID
}
info := AnalyzeInfo{
DBName: as.TableNames[0].Schema.O,
TableName: as.TableNames[0].Name.O,
PartitionName: names[i], TableID: statistics.AnalyzeTableID{TableID: tblInfo.ID, PartitionID: id},
StatsVersion: version,
}
p.ColTasks = append(p.ColTasks, AnalyzeColumnsTask{HandleCols: handleCols, AnalyzeInfo: info, TblInfo: tblInfo})
}
continue
}
}
idx := tblInfo.FindIndexByName(idxName.L)
if idx == nil || idx.State != model.StatePublic {
return nil, plannererrors.ErrAnalyzeMissIndex.GenWithStackByArgs(idxName.O, tblInfo.Name.O)
}
if idx.MVIndex {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing multi-valued indexes is not supported, skip %s", idx.Name.L))
continue
}
if idx.IsColumnarIndex() {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing columnar index is not supported, skip %s", idx.Name.L))
continue
}
p.IdxTasks = append(p.IdxTasks, generateIndexTasks(idx, as, tblInfo, names, physicalIDs, version)...)
}
return p, nil
}
func (b *PlanBuilder) buildAnalyzeAllIndex(as *ast.AnalyzeTableStmt, opts map[ast.AnalyzeOptionType]uint64, version int) (base.Plan, error) {
p := &Analyze{Opts: opts}
statsHandle := domain.GetDomain(b.ctx).StatsHandle()
if statsHandle == nil {
return nil, errors.Errorf("statistics hasn't been initialized, please try again later")
}
tnW := b.resolveCtx.GetTableName(as.TableNames[0])
tblInfo := tnW.TableInfo
physicalIDs, names, err := GetPhysicalIDsAndPartitionNames(tblInfo, as.PartitionNames)
if err != nil {
return nil, err
}
versionIsSame := statsHandle.CheckAnalyzeVersion(tblInfo, physicalIDs, &version)
if !versionIsSame {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("The analyze version from the session is not compatible with the existing statistics of the table. Use the existing version instead"))
}
if version == statistics.Version2 {
return b.buildAnalyzeTable(as, opts, version)
}
for _, idx := range tblInfo.Indices {
if idx.State == model.StatePublic {
if idx.MVIndex {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing multi-valued indexes is not supported, skip %s", idx.Name.L))
continue
}
if idx.IsColumnarIndex() {
b.ctx.GetSessionVars().StmtCtx.AppendWarning(errors.NewNoStackErrorf("analyzing columnar index is not supported, skip %s", idx.Name.L))
continue
}
p.IdxTasks = append(p.IdxTasks, generateIndexTasks(idx, as, tblInfo, names, physicalIDs, version)...)
}
}
handleCols := BuildHandleColsForAnalyze(b.ctx, tblInfo, true, nil)
if handleCols != nil {
for i, id := range physicalIDs {
if id == tblInfo.ID {
id = statistics.NonPartitionTableID
}
info := AnalyzeInfo{
DBName: as.TableNames[0].Schema.O,
TableName: as.TableNames[0].Name.O,
PartitionName: names[i],
TableID: statistics.AnalyzeTableID{TableID: tblInfo.ID, PartitionID: id},
StatsVersion: version,
}
p.ColTasks = append(p.ColTasks, AnalyzeColumnsTask{HandleCols: handleCols, AnalyzeInfo: info, TblInfo: tblInfo})
}
}
return p, nil
}
func generateIndexTasks(idx *model.IndexInfo, as *ast.AnalyzeTableStmt, tblInfo *model.TableInfo, names []string, physicalIDs []int64, version int) []AnalyzeIndexTask {
if idx.Global {
info := AnalyzeInfo{
DBName: as.TableNames[0].Schema.O,
TableName: as.TableNames[0].Name.O,
PartitionName: "",
TableID: statistics.AnalyzeTableID{TableID: tblInfo.ID, PartitionID: statistics.NonPartitionTableID},
StatsVersion: version,
}
return []AnalyzeIndexTask{{IndexInfo: idx, AnalyzeInfo: info, TblInfo: tblInfo}}
}
indexTasks := make([]AnalyzeIndexTask, 0, len(physicalIDs))
for i, id := range physicalIDs {
if id == tblInfo.ID {
id = statistics.NonPartitionTableID
}
info := AnalyzeInfo{
DBName: as.TableNames[0].Schema.O,
TableName: as.TableNames[0].Name.O,
PartitionName: names[i],
TableID: statistics.AnalyzeTableID{TableID: tblInfo.ID, PartitionID: id},
StatsVersion: version,
}
indexTasks = append(indexTasks, AnalyzeIndexTask{IndexInfo: idx, AnalyzeInfo: info, TblInfo: tblInfo})
}
return indexTasks
}
// CMSketchSizeLimit indicates the size limit of CMSketch.
var CMSketchSizeLimit = kv.TxnEntrySizeLimit.Load() / binary.MaxVarintLen32
var analyzeOptionLimit = map[ast.AnalyzeOptionType]uint64{
ast.AnalyzeOptNumBuckets: 100000,
ast.AnalyzeOptNumTopN: 100000,
ast.AnalyzeOptCMSketchWidth: CMSketchSizeLimit,
ast.AnalyzeOptCMSketchDepth: CMSketchSizeLimit,
ast.AnalyzeOptNumSamples: 5000000,
ast.AnalyzeOptSampleRate: math.Float64bits(1),
}
// TODO(0xPoe): give some explanation about the default value.
var analyzeOptionDefault = map[ast.AnalyzeOptionType]uint64{
ast.AnalyzeOptNumBuckets: 256,
ast.AnalyzeOptNumTopN: 20,
ast.AnalyzeOptCMSketchWidth: 2048,
ast.AnalyzeOptCMSketchDepth: 5,
ast.AnalyzeOptNumSamples: 10000,
ast.AnalyzeOptSampleRate: math.Float64bits(0),
}
// TopN reduced from 500 to 100 due to concerns over large number of TopN values collected for customers with many tables.
// 100 is more inline with other databases. 100-256 is also common for NumBuckets with other databases.
var analyzeOptionDefaultV2 = map[ast.AnalyzeOptionType]uint64{
ast.AnalyzeOptNumBuckets: statistics.DefaultHistogramBuckets,
ast.AnalyzeOptNumTopN: statistics.DefaultTopNValue,
ast.AnalyzeOptCMSketchWidth: 2048,
ast.AnalyzeOptCMSketchDepth: 5,
ast.AnalyzeOptNumSamples: 0,
ast.AnalyzeOptSampleRate: math.Float64bits(-1),
}
// GetAnalyzeOptionDefaultV2ForTest returns the default analyze options for test.
func GetAnalyzeOptionDefaultV2ForTest() map[ast.AnalyzeOptionType]uint64 {
return analyzeOptionDefaultV2
}
// This function very similar to handleAnalyzeOptions, but it's used for analyze version 2.
// Remove this function after we remove the support of analyze version 1.
func handleAnalyzeOptionsV2(opts []ast.AnalyzeOpt) (map[ast.AnalyzeOptionType]uint64, error) {
optMap := make(map[ast.AnalyzeOptionType]uint64, len(analyzeOptionDefault))
sampleNum, sampleRate := uint64(0), 0.0
for _, opt := range opts {
datumValue := opt.Value.(*driver.ValueExpr).Datum
switch opt.Type {
case ast.AnalyzeOptNumTopN:
v := datumValue.GetUint64()
if v > analyzeOptionLimit[opt.Type] {
return nil, errors.Errorf("Value of analyze option %s should not be larger than %d", ast.AnalyzeOptionString[opt.Type], analyzeOptionLimit[opt.Type])
}
optMap[opt.Type] = v
case ast.AnalyzeOptSampleRate:
// Only Int/Float/decimal is accepted, so pass nil here is safe.
fVal, err := datumValue.ToFloat64(types.DefaultStmtNoWarningContext)
if err != nil {
return nil, err
}
limit := math.Float64frombits(analyzeOptionLimit[opt.Type])
if fVal <= 0 || fVal > limit {
return nil, errors.Errorf("Value of analyze option %s should not larger than %f, and should be greater than 0", ast.AnalyzeOptionString[opt.Type], limit)
}
sampleRate = fVal
optMap[opt.Type] = math.Float64bits(fVal)
default:
v := datumValue.GetUint64()
if opt.Type == ast.AnalyzeOptNumSamples {
sampleNum = v
}
if v == 0 || v > analyzeOptionLimit[opt.Type] {
return nil, errors.Errorf("Value of analyze option %s should be positive and not larger than %d", ast.AnalyzeOptionString[opt.Type], analyzeOptionLimit[opt.Type])
}
optMap[opt.Type] = v
}
}
if sampleNum > 0 && sampleRate > 0 {
return nil, errors.Errorf("You can only either set the value of the sample num or set the value of the sample rate. Don't set both of them")
}
return optMap, nil
}
func fillAnalyzeOptionsV2(optMap map[ast.AnalyzeOptionType]uint64) map[ast.AnalyzeOptionType]uint64 {
filledMap := make(map[ast.AnalyzeOptionType]uint64, len(analyzeOptionDefault))
for key, defaultVal := range analyzeOptionDefaultV2 {
if val, ok := optMap[key]; ok {
filledMap[key] = val
} else {
filledMap[key] = defaultVal
}
}
return filledMap
}
func handleAnalyzeOptions(opts []ast.AnalyzeOpt, statsVer int) (map[ast.AnalyzeOptionType]uint64, error) {
optMap := make(map[ast.AnalyzeOptionType]uint64, len(analyzeOptionDefault))
if statsVer == statistics.Version1 {
maps.Copy(optMap, analyzeOptionDefault)
} else {
maps.Copy(optMap, analyzeOptionDefaultV2)
}
sampleNum, sampleRate := uint64(0), 0.0
for _, opt := range opts {
datumValue := opt.Value.(*driver.ValueExpr).Datum
switch opt.Type {
case ast.AnalyzeOptNumTopN:
v := datumValue.GetUint64()
if v > analyzeOptionLimit[opt.Type] {
return nil, errors.Errorf("Value of analyze option %s should not be larger than %d", ast.AnalyzeOptionString[opt.Type], analyzeOptionLimit[opt.Type])
}
optMap[opt.Type] = v
case ast.AnalyzeOptSampleRate:
// Only Int/Float/decimal is accepted, so pass nil here is safe.
fVal, err := datumValue.ToFloat64(types.DefaultStmtNoWarningContext)
if err != nil {
return nil, err
}
if fVal > 0 && statsVer == statistics.Version1 {
return nil, errors.Errorf("Version 1's statistics doesn't support the SAMPLERATE option, please set tidb_analyze_version to 2")
}
limit := math.Float64frombits(analyzeOptionLimit[opt.Type])
if fVal <= 0 || fVal > limit {
return nil, errors.Errorf("Value of analyze option %s should not larger than %f, and should be greater than 0", ast.AnalyzeOptionString[opt.Type], limit)
}
sampleRate = fVal
optMap[opt.Type] = math.Float64bits(fVal)
default:
v := datumValue.GetUint64()
if opt.Type == ast.AnalyzeOptNumSamples {
sampleNum = v
}
if v == 0 || v > analyzeOptionLimit[opt.Type] {
return nil, errors.Errorf("Value of analyze option %s should be positive and not larger than %d", ast.AnalyzeOptionString[opt.Type], analyzeOptionLimit[opt.Type])
}
optMap[opt.Type] = v
}
}
if sampleNum > 0 && sampleRate > 0 {
return nil, errors.Errorf("You can only either set the value of the sample num or set the value of the sample rate. Don't set both of them")
}
// Only version 1 has cmsketch.
if statsVer == statistics.Version1 && optMap[ast.AnalyzeOptCMSketchWidth]*optMap[ast.AnalyzeOptCMSketchDepth] > CMSketchSizeLimit {
return nil, errors.Errorf("cm sketch size(depth * width) should not larger than %d", CMSketchSizeLimit)
}
return optMap, nil
}
func (b *PlanBuilder) buildAnalyze(as *ast.AnalyzeTableStmt) (base.Plan, error) {
if as.NoWriteToBinLog {
return nil, dbterror.ErrNotSupportedYet.GenWithStackByArgs("[NO_WRITE_TO_BINLOG | LOCAL]")
}
if as.Incremental {
return nil, errors.Errorf("the incremental analyze feature has already been removed in TiDB v7.5.0, so this will have no effect")
}
statsVersion := b.ctx.GetSessionVars().AnalyzeVersion
// Require INSERT and SELECT privilege for tables.
b.requireInsertAndSelectPriv(as.TableNames)
opts, err := handleAnalyzeOptions(as.AnalyzeOpts, statsVersion)
if err != nil {
return nil, err
}
if as.IndexFlag {
if len(as.IndexNames) == 0 {
return b.buildAnalyzeAllIndex(as, opts, statsVersion)
}
return b.buildAnalyzeIndex(as, opts, statsVersion)
}
return b.buildAnalyzeTable(as, opts, statsVersion)
}
func buildShowNextRowID() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(4)
schema.Append(buildColumnWithName("", "DB_NAME", mysql.TypeVarchar, mysql.MaxDatabaseNameLength))
schema.Append(buildColumnWithName("", "TABLE_NAME", mysql.TypeVarchar, mysql.MaxTableNameLength))
schema.Append(buildColumnWithName("", "COLUMN_NAME", mysql.TypeVarchar, mysql.MaxColumnNameLength))
schema.Append(buildColumnWithName("", "NEXT_GLOBAL_ROW_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "ID_TYPE", mysql.TypeVarchar, 15))
return schema.col2Schema(), schema.names
}
func buildShowDDLFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(6)
schema.Append(buildColumnWithName("", "SCHEMA_VER", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "OWNER_ID", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "OWNER_ADDRESS", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName("", "RUNNING_JOBS", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "SELF_ID", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "QUERY", mysql.TypeVarchar, 256))
return schema.col2Schema(), schema.names
}
func buildRecoverIndexFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(2)
schema.Append(buildColumnWithName("", "ADDED_COUNT", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "SCAN_COUNT", mysql.TypeLonglong, 4))
return schema.col2Schema(), schema.names
}
func buildCleanupIndexFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "REMOVED_COUNT", mysql.TypeLonglong, 4))
return schema.col2Schema(), schema.names
}
func buildShowDDLJobsFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(12)
schema.Append(buildColumnWithName("", "JOB_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "DB_NAME", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "TABLE_NAME", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "JOB_TYPE", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "SCHEMA_STATE", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "SCHEMA_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "TABLE_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "ROW_COUNT", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "CREATE_TIME", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName("", "START_TIME", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName("", "END_TIME", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName("", "STATE", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "COMMENTS", mysql.TypeVarchar, 65535))
return schema.col2Schema(), schema.names
}
func buildTableDistributionSchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(14)
schema.Append(buildColumnWithName("", "PARTITION_NAME", mysql.TypeVarchar, 4))
schema.Append(buildColumnWithName("", "STORE_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "STORE_TYPE", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "REGION_LEADER_COUNT", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_PEER_COUNT", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_WRITE_BYTES", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_WRITE_KEYS", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_WRITE_QUERY", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_LEADER_READ_BYTES", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_LEADER_READ_KEYS", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_LEADER_READ_QUERY", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_PEER_READ_BYTES", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_PEER_READ_KEYS", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "REGION_PEER_READ_QUERY", mysql.TypeLonglong, 4))
return schema.col2Schema(), schema.names
}
func buildTableRegionsSchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(13)
schema.Append(buildColumnWithName("", "REGION_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "START_KEY", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "END_KEY", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "LEADER_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "LEADER_STORE_ID", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "PEERS", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "SCATTERING", mysql.TypeTiny, 1))
schema.Append(buildColumnWithName("", "WRITTEN_BYTES", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "READ_BYTES", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "APPROXIMATE_SIZE(MB)", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "APPROXIMATE_KEYS", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "SCHEDULING_CONSTRAINTS", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "SCHEDULING_STATE", mysql.TypeVarchar, 16))
return schema.col2Schema(), schema.names
}
func buildSplitRegionsSchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(2)
schema.Append(buildColumnWithName("", "TOTAL_SPLIT_REGION", mysql.TypeLonglong, 4))
schema.Append(buildColumnWithName("", "SCATTER_FINISH_RATIO", mysql.TypeDouble, 8))
return schema.col2Schema(), schema.names
}
func buildDistributeTableSchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "JOB_ID", mysql.TypeLonglong, 4))
return schema.col2Schema(), schema.names
}
func buildShowDDLJobQueriesFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "QUERY", mysql.TypeVarchar, 256))
return schema.col2Schema(), schema.names
}
func buildShowDDLJobQueriesWithRangeFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(2)
schema.Append(buildColumnWithName("", "JOB_ID", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "QUERY", mysql.TypeVarchar, 256))
return schema.col2Schema(), schema.names
}
func buildShowSlowSchema() (*expression.Schema, types.NameSlice) {
longlongSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeLonglong)
tinySize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeTiny)
timestampSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeTimestamp)
durationSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeDuration)
schema := newColumnsWithNames(11)
schema.Append(buildColumnWithName("", "SQL", mysql.TypeVarchar, 4096))
schema.Append(buildColumnWithName("", "START", mysql.TypeTimestamp, timestampSize))
schema.Append(buildColumnWithName("", "DURATION", mysql.TypeDuration, durationSize))
schema.Append(buildColumnWithName("", "DETAILS", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "SUCC", mysql.TypeTiny, tinySize))
schema.Append(buildColumnWithName("", "CONN_ID", mysql.TypeLonglong, longlongSize))
schema.Append(buildColumnWithName("", "TRANSACTION_TS", mysql.TypeLonglong, longlongSize))
schema.Append(buildColumnWithName("", "USER", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName("", "DB", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "TABLE_IDS", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "INDEX_IDS", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "INTERNAL", mysql.TypeTiny, tinySize))
schema.Append(buildColumnWithName("", "DIGEST", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "SESSION_ALIAS", mysql.TypeVarchar, 64))
return schema.col2Schema(), schema.names
}
func buildCommandOnDDLJobsFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(2)
schema.Append(buildColumnWithName("", "JOB_ID", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "RESULT", mysql.TypeVarchar, 128))
return schema.col2Schema(), schema.names
}
func buildCancelDDLJobsFields() (*expression.Schema, types.NameSlice) {
return buildCommandOnDDLJobsFields()
}
func buildPauseDDLJobsFields() (*expression.Schema, types.NameSlice) {
return buildCommandOnDDLJobsFields()
}
func buildResumeDDLJobsFields() (*expression.Schema, types.NameSlice) {
return buildCommandOnDDLJobsFields()
}
func buildAdminShowBDRRoleFields() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "BDR_ROLE", mysql.TypeString, 1))
return schema.col2Schema(), schema.names
}
func buildShowBackupMetaSchema() (*expression.Schema, types.NameSlice) {
names := []string{"Database", "Table", "Total_kvs", "Total_bytes", "Time_range_start", "Time_range_end"}
ftypes := []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeDatetime, mysql.TypeDatetime}
schema := newColumnsWithNames(len(names))
for i := range names {
fLen, _ := mysql.GetDefaultFieldLengthAndDecimal(ftypes[i])
if ftypes[i] == mysql.TypeVarchar {
// the default varchar length is `5`, which might be too short for us.
fLen = 255
}
schema.Append(buildColumnWithName("", names[i], ftypes[i], fLen))
}
return schema.col2Schema(), schema.names
}
func buildShowBackupQuerySchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "Query", mysql.TypeVarchar, 4096))
return schema.col2Schema(), schema.names
}
func buildBackupRestoreSchema(kind ast.BRIEKind) (*expression.Schema, types.NameSlice) {
longlongSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeLonglong)
datetimeSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeDatetime)
schema := newColumnsWithNames(5)
schema.Append(buildColumnWithName("", "Destination", mysql.TypeVarchar, 255))
schema.Append(buildColumnWithName("", "Size", mysql.TypeLonglong, longlongSize))
schema.Append(buildColumnWithName("", "BackupTS", mysql.TypeLonglong, longlongSize))
if kind == ast.BRIEKindRestore {
schema.Append(buildColumnWithName("", "Cluster TS", mysql.TypeLonglong, longlongSize))
}
schema.Append(buildColumnWithName("", "Queue Time", mysql.TypeDatetime, datetimeSize))
schema.Append(buildColumnWithName("", "Execution Time", mysql.TypeDatetime, datetimeSize))
return schema.col2Schema(), schema.names
}
func buildBRIESchema(kind ast.BRIEKind) (*expression.Schema, types.NameSlice) {
switch kind {
case ast.BRIEKindShowBackupMeta:
return buildShowBackupMetaSchema()
case ast.BRIEKindShowQuery:
return buildShowBackupQuerySchema()
case ast.BRIEKindBackup, ast.BRIEKindRestore:
return buildBackupRestoreSchema(kind)
default:
s := newColumnsWithNames(0)
return s.col2Schema(), s.names
}
}
func buildCalibrateResourceSchema() (*expression.Schema, types.NameSlice) {
longlongSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeLonglong)
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "QUOTA", mysql.TypeLonglong, longlongSize))
return schema.col2Schema(), schema.names
}
func buildAddQueryWatchSchema() (*expression.Schema, types.NameSlice) {
longlongSize, _ := mysql.GetDefaultFieldLengthAndDecimal(mysql.TypeLonglong)
cols := newColumnsWithNames(1)
cols.Append(buildColumnWithName("", "WATCH_ID", mysql.TypeLonglong, longlongSize))
return cols.col2Schema(), cols.names
}
func buildShowTrafficJobsSchema() (*expression.Schema, types.NameSlice) {
schema := newColumnsWithNames(8)
schema.Append(buildColumnWithName("", "START_TIME", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName("", "END_TIME", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName("", "INSTANCE", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "TYPE", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName("", "PROGRESS", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName("", "STATUS", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName("", "FAIL_REASON", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "PARAMS", mysql.TypeVarchar, 4096))
return schema.col2Schema(), schema.names
}
func buildColumnWithName(tableName, name string, tp byte, size int) (*expression.Column, *types.FieldName) {
cs, cl := types.DefaultCharsetForType(tp)
flag := mysql.UnsignedFlag
if tp == mysql.TypeVarchar || tp == mysql.TypeBlob {
cs = charset.CharsetUTF8MB4
cl = charset.CollationUTF8MB4
flag = 0
}
fieldType := &types.FieldType{}
fieldType.SetType(tp)
fieldType.SetCharset(cs)
fieldType.SetCollate(cl)
fieldType.SetFlen(size)
fieldType.SetFlag(flag)
return &expression.Column{
RetType: fieldType,
}, &types.FieldName{DBName: metadef.InformationSchemaName, TblName: ast.NewCIStr(tableName), ColName: ast.NewCIStr(name)}
}
type columnsWithNames struct {
cols []*expression.Column
names types.NameSlice
}
func newColumnsWithNames(_ int) *columnsWithNames {
return &columnsWithNames{
cols: make([]*expression.Column, 0, 2),
names: make(types.NameSlice, 0, 2),
}
}
func (cwn *columnsWithNames) Append(col *expression.Column, name *types.FieldName) {
cwn.cols = append(cwn.cols, col)
cwn.names = append(cwn.names, name)
}
func (cwn *columnsWithNames) col2Schema() *expression.Schema {
return expression.NewSchema(cwn.cols...)
}
// splitWhere split a where expression to a list of AND conditions.
func splitWhere(where ast.ExprNode) []ast.ExprNode {
var conditions []ast.ExprNode
switch x := where.(type) {
case nil:
case *ast.BinaryOperationExpr:
if x.Op == opcode.LogicAnd {
conditions = append(conditions, splitWhere(x.L)...)
conditions = append(conditions, splitWhere(x.R)...)
} else {
conditions = append(conditions, x)
}
case *ast.ParenthesesExpr:
conditions = append(conditions, splitWhere(x.Expr)...)
default:
conditions = append(conditions, where)
}
return conditions
}
func (b *PlanBuilder) buildShow(ctx context.Context, show *ast.ShowStmt) (base.Plan, error) {
tnW := b.resolveCtx.GetTableName(show.Table)
p := logicalop.LogicalShow{
ShowContents: logicalop.ShowContents{
Tp: show.Tp,
CountWarningsOrErrors: show.CountWarningsOrErrors,
DBName: show.DBName,
Table: tnW,
Partition: show.Partition,
Column: show.Column,
IndexName: show.IndexName,
ResourceGroupName: show.ResourceGroupName,
Flag: show.Flag,
User: show.User,
Roles: show.Roles,
Full: show.Full,
IfNotExists: show.IfNotExists,
GlobalScope: show.GlobalScope,
Extended: show.Extended,
Limit: show.Limit,
ImportJobID: show.ImportJobID,
DistributionJobID: show.DistributionJobID,
ImportGroupKey: show.ShowGroupKey,
},
}.Init(b.ctx)
isView := false
isSequence := false
isTempTableLocal := false
// It depends on ShowPredicateExtractor now
buildPattern := true
switch show.Tp {
case ast.ShowDatabases, ast.ShowVariables, ast.ShowTables, ast.ShowColumns, ast.ShowTableStatus, ast.ShowCollation, ast.ShowAffinity:
if (show.Tp == ast.ShowTables || show.Tp == ast.ShowTableStatus) && p.DBName == "" {
return nil, plannererrors.ErrNoDB
}
if extractor := newShowBaseExtractor(*show); extractor.Extract() {
p.Extractor = extractor
buildPattern = false
}
case ast.ShowCreateTable, ast.ShowCreateSequence, ast.ShowPlacementForTable, ast.ShowPlacementForPartition:
var err error
if table, err := b.is.TableByName(ctx, show.Table.Schema, show.Table.Name); err == nil {
isView = table.Meta().IsView()
isSequence = table.Meta().IsSequence()
isTempTableLocal = table.Meta().TempTableType == model.TempTableLocal
}
user := b.ctx.GetSessionVars().User
if isView {
if user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SHOW VIEW", user.AuthUsername, user.AuthHostname, show.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ShowViewPriv, show.Table.Schema.L, show.Table.Name.L, "", err)
} else {
if user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SHOW", user.AuthUsername, user.AuthHostname, show.Table.Name.L)
}
if isTempTableLocal {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AllPrivMask, show.Table.Schema.L, show.Table.Name.L, "", err)
} else {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AllPrivMask&(^mysql.CreateTMPTablePriv), show.Table.Schema.L, show.Table.Name.L, "", err)
}
}
case ast.ShowConfig:
privErr := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("CONFIG")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ConfigPriv, "", "", "", privErr)
case ast.ShowCreateView:
var err error
user := b.ctx.GetSessionVars().User
if user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SELECT", user.AuthUsername, user.AuthHostname, show.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, show.Table.Schema.L, show.Table.Name.L, "", err)
if user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SHOW VIEW", user.AuthUsername, user.AuthHostname, show.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ShowViewPriv, show.Table.Schema.L, show.Table.Name.L, "", err)
case ast.ShowBackups:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or BACKUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"BACKUP_ADMIN"}, false, err)
case ast.ShowRestores:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESTORE_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESTORE_ADMIN"}, false, err)
case ast.ShowTableNextRowId:
p := &ShowNextRowID{TableName: show.Table}
p.SetSchemaAndNames(buildShowNextRowID())
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, show.Table.Schema.L, show.Table.Name.L, "", plannererrors.ErrPrivilegeCheckFail)
return p, nil
case ast.ShowStatsExtended, ast.ShowStatsHealthy, ast.ShowStatsTopN, ast.ShowHistogramsInFlight, ast.ShowColumnStatsUsage:
var err error
if user := b.ctx.GetSessionVars().User; user != nil {
err = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(user.AuthUsername, user.AuthHostname, mysql.SystemDB)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, mysql.SystemDB, "", "", err)
if show.Tp == ast.ShowStatsHealthy {
if extractor := newShowBaseExtractor(*show); extractor.Extract() {
p.Extractor = extractor
buildPattern = false
}
}
case ast.ShowStatsBuckets, ast.ShowStatsHistograms, ast.ShowStatsMeta, ast.ShowStatsLocked:
var err error
if user := b.ctx.GetSessionVars().User; user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SHOW", user.AuthUsername, user.AuthHostname, show.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, show.Table.Schema.L, show.Table.Name.L, "", err)
if show.Tp == ast.ShowStatsMeta {
if extractor := newShowBaseExtractor(*show); extractor.Extract() {
p.Extractor = extractor
buildPattern = false
}
}
case ast.ShowRegions:
tableInfo, err := b.is.TableByName(ctx, show.Table.Schema, show.Table.Name)
if err != nil {
return nil, err
}
if tableInfo.Meta().TempTableType != model.TempTableNone {
return nil, plannererrors.ErrOptOnTemporaryTable.GenWithStackByArgs("show table regions")
}
case ast.ShowDistributions:
tableInfo, err := b.is.TableByName(ctx, show.Table.Schema, show.Table.Name)
if err != nil {
return nil, err
}
if tableInfo.Meta().TempTableType != model.TempTableNone {
return nil, plannererrors.ErrOptOnTemporaryTable.GenWithStackByArgs("show table distributions")
}
case ast.ShowReplicaStatus:
return nil, dbterror.ErrNotSupportedYet.GenWithStackByArgs("SHOW {REPLICA | SLAVE} STATUS")
}
schema, names := buildShowSchema(show, isView, isSequence)
p.SetSchema(schema)
p.SetOutputNames(names)
for _, col := range p.Schema().Columns {
col.UniqueID = b.ctx.GetSessionVars().AllocPlanColumnID()
}
var err error
var np base.LogicalPlan
np = p
// If we have ShowPredicateExtractor, we do not buildSelection with Pattern
if show.Pattern != nil && buildPattern {
show.Pattern.Expr = &ast.ColumnNameExpr{
Name: &ast.ColumnName{Name: p.OutputNames()[0].ColName},
}
np, err = b.buildSelection(ctx, np, show.Pattern, nil)
if err != nil {
return nil, err
}
}
if show.Where != nil {
np, err = b.buildSelection(ctx, np, show.Where, nil)
if err != nil {
return nil, err
}
}
if show.Limit != nil {
np, err = b.buildLimit(np, show.Limit)
if err != nil {
return nil, err
}
}
if np != p {
b.optFlag |= rule.FlagEliminateProjection
fieldsLen := len(p.Schema().Columns)
proj := logicalop.LogicalProjection{Exprs: make([]expression.Expression, 0, fieldsLen)}.Init(b.ctx, 0)
schema := expression.NewSchema(make([]*expression.Column, 0, fieldsLen)...)
for _, col := range p.Schema().Columns {
proj.Exprs = append(proj.Exprs, col)
newCol := col.Clone().(*expression.Column)
newCol.UniqueID = b.ctx.GetSessionVars().AllocPlanColumnID()
schema.Append(newCol)
}
proj.SetSchema(schema)
proj.SetChildren(np)
proj.SetOutputNames(np.OutputNames())
np = proj
}
if show.Tp == ast.ShowVariables || show.Tp == ast.ShowStatus {
b.curClause = orderByClause
orderByCol := np.Schema().Columns[0].Clone().(*expression.Column)
sort := logicalop.LogicalSort{
ByItems: []*util.ByItems{{Expr: orderByCol}},
}.Init(b.ctx, b.getSelectOffset())
sort.SetChildren(np)
np = sort
}
return np, nil
}
func (b *PlanBuilder) buildSimple(ctx context.Context, node ast.StmtNode) (base.Plan, error) {
p := &Simple{Statement: node, ResolveCtx: b.resolveCtx}
switch raw := node.(type) {
case *ast.FlushStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("RELOAD")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ReloadPriv, "", "", "", err)
case *ast.AlterInstanceStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", err)
case *ast.RenameUserStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("CREATE USER")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreateUserPriv, "", "", "", err)
case *ast.GrantStmt:
var err error
b.visitInfo, err = collectVisitInfoFromGrantStmt(b.ctx, b.visitInfo, raw)
if err != nil {
return nil, err
}
for _, user := range raw.Users {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, user.User, "RESTRICTED_USER_ADMIN")
}
case *ast.BRIEStmt:
p.SetSchemaAndNames(buildBRIESchema(raw.Kind))
if raw.Kind == ast.BRIEKindRestore {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESTORE_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESTORE_ADMIN"}, false, err)
} else {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or BACKUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"BACKUP_ADMIN"}, false, err)
}
case *ast.CalibrateResourceStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESOURCE_GROUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESOURCE_GROUP_ADMIN"}, false, err)
p.SetSchemaAndNames(buildCalibrateResourceSchema())
case *ast.AddQueryWatchStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESOURCE_GROUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESOURCE_GROUP_ADMIN"}, false, err)
p.SetSchemaAndNames(buildAddQueryWatchSchema())
case *ast.DropQueryWatchStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESOURCE_GROUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESOURCE_GROUP_ADMIN"}, false, err)
case *ast.GrantRoleStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or ROLE_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"ROLE_ADMIN"}, false, err)
// If any of the roles are RESTRICTED, require RESTRICTED_ROLE_ADMIN
for _, role := range raw.Roles {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, &auth.UserIdentity{Username: role.Username, Hostname: role.Hostname},
"RESTRICTED_USER_ADMIN")
}
case *ast.RevokeRoleStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or ROLE_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"ROLE_ADMIN"}, false, err)
// Check if any of the users are RESTRICTED
for _, user := range raw.Users {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, user, "RESTRICTED_USER_ADMIN")
}
// If any of the roles are RESTRICTED, require RESTRICTED_ROLE_ADMIN
for _, role := range raw.Roles {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, &auth.UserIdentity{Username: role.Username, Hostname: role.Hostname},
"RESTRICTED_USER_ADMIN")
}
case *ast.RevokeStmt:
var err error
b.visitInfo, err = collectVisitInfoFromRevokeStmt(ctx, b.ctx, b.visitInfo, raw)
if err != nil {
return nil, err
}
for _, user := range raw.Users {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, user.User, "RESTRICTED_USER_ADMIN")
}
case *ast.KillStmt:
// All users can kill their own connections regardless.
// If you have the SUPER privilege, you can kill all threads and statements unless SEM is enabled.
// In which case you require RESTRICTED_CONNECTION_ADMIN to kill connections that belong to RESTRICTED_USER_ADMIN users.
sm := b.ctx.GetSessionManager()
if sm != nil {
if pi, ok := sm.GetProcessInfo(raw.ConnectionID); ok {
loginUser := b.ctx.GetSessionVars().User
if pi.User != loginUser.Username {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or CONNECTION_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"CONNECTION_ADMIN"}, false, err)
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, &auth.UserIdentity{Username: pi.User, Hostname: pi.Host}, "RESTRICTED_CONNECTION_ADMIN")
}
} else if handleutil.GlobalAutoAnalyzeProcessList.Contains(raw.ConnectionID) {
// Only the users with SUPER or CONNECTION_ADMIN privilege can kill auto analyze.
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or CONNECTION_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"CONNECTION_ADMIN"}, false, err)
}
}
case *ast.UseStmt:
if raw.DBName == "" {
return nil, plannererrors.ErrNoDB
}
case *ast.DropUserStmt:
// The main privilege checks for DROP USER are currently performed in executor/simple.go
// because they use complex OR conditions (not supported by visitInfo).
for _, user := range raw.UserList {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, user, "RESTRICTED_USER_ADMIN")
}
case *ast.SetPwdStmt:
if raw.User != nil {
b.visitInfo = appendVisitInfoIsRestrictedUser(ctx, b.visitInfo, b.ctx, raw.User, "RESTRICTED_USER_ADMIN")
}
case *ast.ShutdownStmt:
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ShutdownPriv, "", "", "", nil)
case *ast.BeginStmt:
readTS := b.ctx.GetSessionVars().TxnReadTS.PeakTxnReadTS()
if raw.AsOf != nil {
startTS, err := staleread.CalculateAsOfTsExpr(ctx, b.ctx, raw.AsOf.TsExpr)
if err != nil {
return nil, err
}
if err := sessionctx.ValidateSnapshotReadTS(ctx, b.ctx.GetStore(), startTS, true); err != nil {
return nil, err
}
p.StaleTxnStartTS = startTS
} else if readTS > 0 {
p.StaleTxnStartTS = readTS
// consume read ts here
b.ctx.GetSessionVars().TxnReadTS.UseTxnReadTS()
} else if b.ctx.GetSessionVars().EnableExternalTSRead && !b.ctx.GetSessionVars().InRestrictedSQL {
// try to get the stale ts from external timestamp
startTS, err := staleread.GetExternalTimestamp(ctx, b.ctx.GetSessionVars().StmtCtx)
if err != nil {
return nil, err
}
if err := sessionctx.ValidateSnapshotReadTS(ctx, b.ctx.GetStore(), startTS, true); err != nil {
return nil, err
}
p.StaleTxnStartTS = startTS
}
case *ast.SetResourceGroupStmt:
if vardef.EnableResourceControlStrictMode.Load() {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESOURCE_GROUP_ADMIN or RESOURCE_GROUP_USER")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESOURCE_GROUP_ADMIN", "RESOURCE_GROUP_USER"}, false, err)
}
case *ast.DropProcedureStmt:
procName := fmt.Sprintf("%s.%s", ast.NewCIStr(b.ctx.GetSessionVars().CurrentDB).O, raw.ProcedureName.Name.O)
err := plannererrors.ErrSpDoesNotExist.GenWithStackByArgs("PROCEDURE", procName)
if !raw.IfExists {
return nil, err
}
b.ctx.GetSessionVars().StmtCtx.AppendNote(err)
}
return p, nil
}
func collectVisitInfoFromRevokeStmt(ctx context.Context, sctx base.PlanContext, vi []visitInfo, stmt *ast.RevokeStmt) ([]visitInfo, error) {
// To use REVOKE, you must have the GRANT OPTION privilege,
// and you must have the privileges that you are granting.
dbName := stmt.Level.DBName
tableName := stmt.Level.TableName
// This supports a local revoke SELECT on tablename, but does
// not add dbName to the visitInfo of a *.* grant.
if dbName == "" && stmt.Level.Level != ast.GrantLevelGlobal {
if sctx.GetSessionVars().CurrentDB == "" {
return nil, plannererrors.ErrNoDB
}
dbName = strings.ToLower(sctx.GetSessionVars().CurrentDB)
}
var nonDynamicPrivilege bool
var allPrivs []mysql.PrivilegeType
for _, item := range stmt.Privs {
if semv2.IsEnabled() {
if (len(item.Name) > 0 && semv2.IsRestrictedPrivilege(strings.ToUpper(item.Name))) ||
(len(item.Name) == 0 && semv2.IsRestrictedPrivilege(strings.ToUpper(item.Priv.String()))) {
// In `semv2`, we'll support to limit non-dynamic privileges unless the user has the `RESTRICTED_PRIV_ADMIN` privilege.
// For example, `File` privilege might be restricted.
// It's also controlled by the `GRANT OPTION`, so the user will also need the `GRANT OPTION` for this privilege.
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("RESTRICTED_PRIV_ADMIN")
vi = appendDynamicVisitInfo(vi, []string{"RESTRICTED_PRIV_ADMIN"}, false, err)
}
}
if item.Priv == mysql.ExtendedPriv {
vi = appendDynamicVisitInfo(vi, []string{strings.ToUpper(item.Name)}, true, nil) // verified in MySQL: requires the dynamic grant option to revoke.
continue
}
nonDynamicPrivilege = true
if item.Priv == mysql.AllPriv {
switch stmt.Level.Level {
case ast.GrantLevelGlobal:
allPrivs = mysql.AllGlobalPrivs
case ast.GrantLevelDB:
allPrivs = mysql.AllDBPrivs
case ast.GrantLevelTable:
allPrivs = mysql.AllTablePrivs
}
break
}
vi = appendVisitInfo(vi, item.Priv, dbName, tableName, "", nil)
}
for _, priv := range allPrivs {
vi = appendVisitInfo(vi, priv, dbName, tableName, "", nil)
}
for _, u := range stmt.Users {
// For SEM, make sure the users are not restricted
vi = appendVisitInfoIsRestrictedUser(ctx, vi, sctx, u.User, "RESTRICTED_USER_ADMIN")
}
if nonDynamicPrivilege {
// Dynamic privileges use their own GRANT OPTION. If there were any non-dynamic privilege requests,
// we need to attach the "GLOBAL" version of the GRANT OPTION.
vi = appendVisitInfo(vi, mysql.GrantPriv, dbName, tableName, "", nil)
}
return vi, nil
}
// appendVisitInfoIsRestrictedUser appends additional visitInfo if the user has a
// special privilege called "RESTRICTED_USER_ADMIN". It only applies when SEM is enabled.
func appendVisitInfoIsRestrictedUser(ctx context.Context, visitInfo []visitInfo, sctx base.PlanContext, user *auth.UserIdentity, priv string) []visitInfo {
if !sem.IsEnabled() {
return visitInfo
}
checker := privilege.GetPrivilegeManager(sctx)
if checker != nil && checker.RequestDynamicVerificationWithUser(ctx, "RESTRICTED_USER_ADMIN", false, user) {
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs(priv)
visitInfo = appendDynamicVisitInfo(visitInfo, []string{priv}, false, err)
}
return visitInfo
}
func collectVisitInfoFromGrantStmt(sctx base.PlanContext, vi []visitInfo, stmt *ast.GrantStmt) ([]visitInfo, error) {
// To use GRANT, you must have the GRANT OPTION privilege,
// and you must have the privileges that you are granting.
dbName := stmt.Level.DBName
tableName := stmt.Level.TableName
// This supports a local revoke SELECT on tablename, but does
// not add dbName to the visitInfo of a *.* grant.
if dbName == "" && stmt.Level.Level != ast.GrantLevelGlobal {
if sctx.GetSessionVars().CurrentDB == "" {
return nil, plannererrors.ErrNoDB
}
dbName = sctx.GetSessionVars().CurrentDB
}
var nonDynamicPrivilege bool
var allPrivs []mysql.PrivilegeType
authErr := genAuthErrForGrantStmt(sctx, dbName)
for _, item := range stmt.Privs {
if semv2.IsEnabled() {
if (len(item.Name) > 0 && semv2.IsRestrictedPrivilege(strings.ToUpper(item.Name))) ||
(len(item.Name) == 0 && semv2.IsRestrictedPrivilege(strings.ToUpper(item.Priv.String()))) {
// In `semv2`, we'll support to limit non-dynamic privileges unless the user has the `RESTRICTED_PRIV_ADMIN` privilege.
// For example, `File` privilege might be restricted.
// It's also controlled by the `GRANT OPTION`, so the user will also need the `GRANT OPTION` for this privilege.
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("RESTRICTED_PRIV_ADMIN")
vi = appendDynamicVisitInfo(vi, []string{"RESTRICTED_PRIV_ADMIN"}, false, err)
}
}
if item.Priv == mysql.ExtendedPriv {
// The observed MySQL behavior is that the error is:
// ERROR 1227 (42000): Access denied; you need (at least one of) the GRANT OPTION privilege(s) for this operation
// This is ambiguous, because it doesn't say the GRANT OPTION for which dynamic privilege.
// In privilege/privileges/cache.go:RequestDynamicVerification SUPER+Grant_Priv will also be accepted here by TiDB, but it is *not* by MySQL.
// This extension is currently required because:
// - The visitInfo system does not accept OR conditions. There are many scenarios where SUPER or a DYNAMIC privilege are supported,
// this is the one case where SUPER is not intended to be an alternative.
// - The "ALL" privilege for TiDB does not include all dynamic privileges. This could be fixed by a bootstrap task to assign all SUPER users
// with dynamic privileges.
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("GRANT OPTION")
vi = appendDynamicVisitInfo(vi, []string{item.Name}, true, err)
continue
}
nonDynamicPrivilege = true
if item.Priv == mysql.AllPriv {
switch stmt.Level.Level {
case ast.GrantLevelGlobal:
allPrivs = mysql.AllGlobalPrivs
case ast.GrantLevelDB:
allPrivs = mysql.AllDBPrivs
case ast.GrantLevelTable:
allPrivs = mysql.AllTablePrivs
}
break
}
vi = appendVisitInfo(vi, item.Priv, dbName, tableName, "", authErr)
}
for _, priv := range allPrivs {
vi = appendVisitInfo(vi, priv, dbName, tableName, "", authErr)
}
if nonDynamicPrivilege {
// Dynamic privileges use their own GRANT OPTION. If there were any non-dynamic privilege requests,
// we need to attach the "GLOBAL" version of the GRANT OPTION.
vi = appendVisitInfo(vi, mysql.GrantPriv, dbName, tableName, "", authErr)
}
return vi, nil
}
func genAuthErrForGrantStmt(sctx base.PlanContext, dbName string) error {
if !strings.EqualFold(dbName, metadef.PerformanceSchemaName.L) {
return nil
}
user := sctx.GetSessionVars().User
if user == nil {
return nil
}
u := user.Username
h := user.Hostname
if len(user.AuthUsername) > 0 && len(user.AuthHostname) > 0 {
u = user.AuthUsername
h = user.AuthHostname
}
return plannererrors.ErrDBaccessDenied.FastGenByArgs(u, h, dbName)
}
func (b *PlanBuilder) getDefaultValueForInsert(col *table.Column) (*expression.Constant, error) {
var (
value types.Datum
err error
)
if col.DefaultIsExpr && col.DefaultExpr != nil {
value, err = table.EvalColDefaultExpr(b.ctx.GetExprCtx(), col.ToInfo(), col.DefaultExpr)
} else {
if err := table.CheckNoDefaultValueForInsert(b.ctx.GetSessionVars().StmtCtx, col.ToInfo()); err != nil {
return nil, err
}
value, err = table.GetColDefaultValue(b.ctx.GetExprCtx(), col.ToInfo())
}
if err != nil {
return nil, err
}
return &expression.Constant{Value: value, RetType: col.FieldType.Clone()}, nil
}
// resolveGeneratedColumns resolves generated columns with their generation expressions respectively.
// If not-nil onDups is passed in, it will be **modified in place** and record columns in on-duplicate list
func (b *PlanBuilder) resolveGeneratedColumns(ctx context.Context, columns []*table.Column, onDups map[string]struct{}, mockPlan base.LogicalPlan) (igc physicalop.InsertGeneratedColumns, err error) {
for _, column := range columns {
if !column.IsGenerated() {
// columns having on-update-now flag should also be considered.
if onDups != nil && mysql.HasOnUpdateNowFlag(column.GetFlag()) {
onDups[column.Name.L] = struct{}{}
}
continue
}
columnName := &ast.ColumnName{Name: column.Name}
columnName.SetText(nil, column.Name.O)
idx, err := expression.FindFieldName(mockPlan.OutputNames(), columnName)
if err != nil {
return igc, err
}
colExpr := mockPlan.Schema().Columns[idx]
originalVal := b.allowBuildCastArray
b.allowBuildCastArray = true
expr, _, err := b.rewrite(ctx, column.GeneratedExpr.Clone(), mockPlan, nil, true)
b.allowBuildCastArray = originalVal
if err != nil {
return igc, err
}
igc.Exprs = append(igc.Exprs, expr)
if onDups == nil {
continue
}
// There may be chain dependencies between columns,
// so we need to add new columns into onDups.
for dep := range column.Dependences {
if _, ok := onDups[dep]; ok {
assign := &expression.Assignment{Col: colExpr, ColName: column.Name, Expr: expr}
igc.OnDuplicates = append(igc.OnDuplicates, assign)
// onDups use lower column name, see Insert.resolveOnDuplicate
onDups[column.Name.L] = struct{}{}
break
}
}
}
return igc, nil
}
func (b *PlanBuilder) buildInsert(ctx context.Context, insert *ast.InsertStmt) (base.Plan, error) {
ts, ok := insert.Table.TableRefs.Left.(*ast.TableSource)
if !ok {
return nil, infoschema.ErrTableNotExists.FastGenByArgs()
}
tn, ok := ts.Source.(*ast.TableName)
if !ok {
return nil, infoschema.ErrTableNotExists.FastGenByArgs()
}
tnW := b.resolveCtx.GetTableName(tn)
tableInfo := tnW.TableInfo
if tableInfo.IsView() {
err := errors.Errorf("insert into view %s is not supported now", tableInfo.Name.O)
if insert.IsReplace {
err = errors.Errorf("replace into view %s is not supported now", tableInfo.Name.O)
}
return nil, err
}
if tableInfo.IsSequence() {
err := errors.Errorf("insert into sequence %s is not supported now", tableInfo.Name.O)
if insert.IsReplace {
err = errors.Errorf("replace into sequence %s is not supported now", tableInfo.Name.O)
}
return nil, err
}
// Build Schema with DBName otherwise ColumnRef with DBName cannot match any Column in Schema.
schema, names, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), tn.Schema, tableInfo)
if err != nil {
return nil, err
}
tableInPlan, ok := b.is.TableByID(ctx, tableInfo.ID)
if !ok {
return nil, errors.Errorf("Can't get table %s", tableInfo.Name.O)
}
insertPlan := physicalop.Insert{
Table: tableInPlan,
Columns: insert.Columns,
TableSchema: schema,
TableColNames: names,
IsReplace: insert.IsReplace,
IgnoreErr: insert.IgnoreErr,
}.Init(b.ctx)
if tableInfo.GetPartitionInfo() != nil && len(insert.PartitionNames) != 0 {
givenPartitionSets := make(map[int64]struct{}, len(insert.PartitionNames))
// check partition by name.
for _, name := range insert.PartitionNames {
id, err := tables.FindPartitionByName(tableInfo, name.L)
if err != nil {
return nil, err
}
givenPartitionSets[id] = struct{}{}
}
pt := tableInPlan.(table.PartitionedTable)
insertPlan.Table = tables.NewPartitionTableWithGivenSets(pt, givenPartitionSets)
} else if len(insert.PartitionNames) != 0 {
return nil, plannererrors.ErrPartitionClauseOnNonpartitioned
}
user := b.ctx.GetSessionVars().User
var authErr error
if user != nil {
authErr = plannererrors.ErrTableaccessDenied.FastGenByArgs("INSERT", user.AuthUsername, user.AuthHostname, tableInfo.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, tnW.DBInfo.Name.L,
tableInfo.Name.L, "", authErr)
// `REPLACE INTO` requires both INSERT + DELETE privilege
// `ON DUPLICATE KEY UPDATE` requires both INSERT + UPDATE privilege
var extraPriv mysql.PrivilegeType
if insert.IsReplace {
extraPriv = mysql.DeletePriv
} else if insert.OnDuplicate != nil {
extraPriv = mysql.UpdatePriv
}
if extraPriv != 0 {
if user != nil {
cmd := strings.ToUpper(mysql.Priv2Str[extraPriv])
authErr = plannererrors.ErrTableaccessDenied.FastGenByArgs(cmd, user.AuthUsername, user.AuthHostname, tableInfo.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, extraPriv, tnW.DBInfo.Name.L, tableInfo.Name.L, "", authErr)
}
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
mockTablePlan.SetSchema(insertPlan.TableSchema)
mockTablePlan.SetOutputNames(insertPlan.TableColNames)
checkRefColumn := func(n ast.Node) ast.Node {
if insertPlan.NeedFillDefaultValue {
return n
}
switch n.(type) {
case *ast.ColumnName, *ast.ColumnNameExpr:
insertPlan.NeedFillDefaultValue = true
}
return n
}
if len(insert.Lists) > 0 {
// Branch for `INSERT ... VALUES ...`.
// Branch for `INSERT ... SET ...`.
err := b.buildValuesListOfInsert(ctx, insert, insertPlan, mockTablePlan, checkRefColumn)
if err != nil {
return nil, err
}
} else {
// Branch for `INSERT ... SELECT ...`.
err := b.buildSelectPlanOfInsert(ctx, insert, insertPlan)
if err != nil {
return nil, err
}
}
mockTablePlan.SetSchema(insertPlan.Schema4OnDuplicate)
mockTablePlan.SetOutputNames(insertPlan.Names4OnDuplicate)
onDupColSet, err := insertPlan.ResolveOnDuplicate(insert.OnDuplicate, tableInfo, func(node ast.ExprNode) (expression.Expression, error) {
return b.rewriteInsertOnDuplicateUpdate(ctx, node, mockTablePlan, insertPlan)
})
if err != nil {
return nil, err
}
// Calculate generated columns.
mockTablePlan.SetSchema(insertPlan.TableSchema)
mockTablePlan.SetOutputNames(insertPlan.TableColNames)
insertPlan.GenCols, err = b.resolveGeneratedColumns(ctx, insertPlan.Table.Cols(), onDupColSet, mockTablePlan)
if err != nil {
return nil, err
}
err = insertPlan.ResolveIndices()
if err != nil {
return nil, err
}
err = insertPlan.BuildOnInsertFKTriggers(b.ctx, b.is, tnW.DBInfo.Name.L)
return insertPlan, err
}
func (*PlanBuilder) getAffectCols(insertStmt *ast.InsertStmt, insertPlan *physicalop.Insert) (affectedValuesCols []*table.Column, err error) {
if len(insertStmt.Columns) > 0 {
// This branch is for the following scenarios:
// 1. `INSERT INTO tbl_name (col_name [, col_name] ...) {VALUES | VALUE} (value_list) [, (value_list)] ...`,
// 2. `INSERT INTO tbl_name (col_name [, col_name] ...) SELECT ...`.
// 3. `INSERT INTO tbl_name SET col1=x1, ... colM=xM...`.
colName := make([]string, 0, len(insertStmt.Columns))
for _, col := range insertStmt.Columns {
colName = append(colName, col.Name.L)
}
var missingColIdx int
affectedValuesCols, missingColIdx = table.FindColumns(insertPlan.Table.VisibleCols(), colName, insertPlan.Table.Meta().PKIsHandle)
if missingColIdx >= 0 {
return nil, plannererrors.ErrUnknownColumn.GenWithStackByArgs(
insertStmt.Columns[missingColIdx].Name.O, clauseMsg[fieldList])
}
} else {
// This branch is for the following scenarios:
// 1. `INSERT INTO tbl_name {VALUES | VALUE} (value_list) [, (value_list)] ...`,
// 2. `INSERT INTO tbl_name SELECT ...`.
affectedValuesCols = insertPlan.Table.VisibleCols()
}
return affectedValuesCols, nil
}
func (b PlanBuilder) getInsertColExpr(ctx context.Context, insertPlan *physicalop.Insert, mockTablePlan *logicalop.LogicalTableDual, col *table.Column, expr ast.ExprNode, checkRefColumn func(n ast.Node) ast.Node) (outExpr expression.Expression, err error) {
if col.Hidden {
return nil, plannererrors.ErrUnknownColumn.GenWithStackByArgs(col.Name, clauseMsg[fieldList])
}
switch x := expr.(type) {
case *ast.DefaultExpr:
refCol := col
if x.Name != nil {
refCol = table.FindColLowerCase(insertPlan.Table.Cols(), x.Name.Name.L)
if refCol == nil {
return nil, plannererrors.ErrUnknownColumn.GenWithStackByArgs(x.Name.OrigColName(), clauseMsg[fieldList])
}
// Cannot use DEFAULT(generated column) except for the same column
if col != refCol && (col.IsGenerated() || refCol.IsGenerated()) {
return nil, plannererrors.ErrBadGeneratedColumn.GenWithStackByArgs(col.Name.O, insertPlan.Table.Meta().Name.O)
} else if col == refCol && col.IsGenerated() {
return nil, nil
}
} else if col.IsGenerated() {
// See note in the end of the function. Only default for generated columns are OK.
return nil, nil
}
outExpr, err = b.getDefaultValueForInsert(refCol)
case *driver.ValueExpr:
outExpr = &expression.Constant{
Value: x.Datum,
RetType: &x.Type,
}
case *driver.ParamMarkerExpr:
outExpr, err = expression.ParamMarkerExpression(b.ctx.GetExprCtx(), x, false)
default:
b.curClause = fieldList
// subquery in insert values should not reference upper scope
usingPlan := mockTablePlan
if _, ok := expr.(*ast.SubqueryExpr); ok {
usingPlan = logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
}
var np base.LogicalPlan
outExpr, np, err = b.rewriteWithPreprocess(ctx, expr, usingPlan, nil, nil, true, checkRefColumn)
if np != nil {
if _, ok := np.(*logicalop.LogicalTableDual); !ok {
// See issue#30626 and the related tests in function TestInsertValuesWithSubQuery for more details.
// This is a TODO and we will support it later.
return nil, errors.New("Insert's SET operation or VALUES_LIST doesn't support complex subqueries now")
}
}
}
if err != nil {
return nil, err
}
if insertPlan.AllAssignmentsAreConstant {
_, isConstant := outExpr.(*expression.Constant)
insertPlan.AllAssignmentsAreConstant = isConstant
}
// Note: For INSERT, REPLACE, and UPDATE, if a generated column is inserted into, replaced, or updated explicitly, the only permitted value is DEFAULT.
// see https://dev.mysql.com/doc/refman/8.0/en/create-table-generated-columns.html
if col.IsGenerated() {
return nil, plannererrors.ErrBadGeneratedColumn.GenWithStackByArgs(col.Name.O, insertPlan.Table.Meta().Name.O)
}
return outExpr, nil
}
func (b *PlanBuilder) buildValuesListOfInsert(ctx context.Context, insert *ast.InsertStmt, insertPlan *physicalop.Insert, mockTablePlan *logicalop.LogicalTableDual, checkRefColumn func(n ast.Node) ast.Node) error {
affectedValuesCols, err := b.getAffectCols(insert, insertPlan)
if err != nil {
return err
}
// If value_list and col_list are empty and we have a generated column, we can still write data to this table.
// For example, insert into t values(); can be executed successfully if t has a generated column.
if len(insert.Columns) > 0 || len(insert.Lists[0]) > 0 {
// If value_list or col_list is not empty, the length of value_list should be the same with that of col_list.
if len(insert.Lists[0]) != len(affectedValuesCols) {
return plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(1)
}
}
insertPlan.AllAssignmentsAreConstant = true
for i, valuesItem := range insert.Lists {
// The length of all the value_list should be the same.
// "insert into t values (), ()" is valid.
// "insert into t values (), (1)" is not valid.
// "insert into t values (1), ()" is not valid.
// "insert into t values (1,2), (1)" is not valid.
if i > 0 && len(insert.Lists[i-1]) != len(insert.Lists[i]) {
return plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(i + 1)
}
exprList := make([]expression.Expression, 0, len(valuesItem))
for j, valueItem := range valuesItem {
expr, err := b.getInsertColExpr(ctx, insertPlan, mockTablePlan, affectedValuesCols[j], valueItem, checkRefColumn)
if err != nil {
return err
}
if expr == nil {
continue
}
exprList = append(exprList, expr)
}
insertPlan.Lists = append(insertPlan.Lists, exprList)
}
insertPlan.Schema4OnDuplicate = insertPlan.TableSchema
insertPlan.Names4OnDuplicate = insertPlan.TableColNames
return nil
}
type colNameInOnDupExtractor struct {
colNameMap map[types.FieldName]*ast.ColumnNameExpr
}
func (c *colNameInOnDupExtractor) Enter(node ast.Node) (ast.Node, bool) {
switch x := node.(type) {
case *ast.ColumnNameExpr:
fieldName := types.FieldName{
DBName: x.Name.Schema,
TblName: x.Name.Table,
ColName: x.Name.Name,
}
c.colNameMap[fieldName] = x
return node, true
// We don't extract the column names from the sub select.
case *ast.SelectStmt, *ast.SetOprStmt:
return node, true
default:
return node, false
}
}
func (*colNameInOnDupExtractor) Leave(node ast.Node) (ast.Node, bool) {
return node, true
}
func (b *PlanBuilder) buildSelectPlanOfInsert(ctx context.Context, insert *ast.InsertStmt, insertPlan *physicalop.Insert) error {
b.isForUpdateRead = true
affectedValuesCols, err := b.getAffectCols(insert, insertPlan)
if err != nil {
return err
}
actualColLen := -1
// For MYSQL, it handles the case that the columns in ON DUPLICATE UPDATE is not the project column of the SELECT clause
// but just in the table occurs in the SELECT CLAUSE.
// e.g. insert into a select x from b ON DUPLICATE KEY UPDATE a.x=b.y; the `y` is not a column of select's output.
// MySQL won't throw error and will execute this SQL successfully.
// To make compatible with this strange feature, we add the variable `actualColLen` and the following IF branch.
if len(insert.OnDuplicate) > 0 {
// If the select has aggregation, it cannot see the columns not in the select field.
// e.g. insert into a select x from b ON DUPLICATE KEY UPDATE a.x=b.y; can be executed successfully.
// insert into a select x from b group by x ON DUPLICATE KEY UPDATE a.x=b.y; will report b.y not found.
if sel, ok := insert.Select.(*ast.SelectStmt); ok && !b.detectSelectAgg(sel) {
hasWildCard := false
for _, field := range sel.Fields.Fields {
if field.WildCard != nil {
hasWildCard = true
break
}
}
if !hasWildCard {
colExtractor := &colNameInOnDupExtractor{colNameMap: make(map[types.FieldName]*ast.ColumnNameExpr)}
for _, assign := range insert.OnDuplicate {
assign.Expr.Accept(colExtractor)
}
actualColLen = len(sel.Fields.Fields)
for _, colName := range colExtractor.colNameMap {
// If we found the name from the INSERT's table columns, we don't try to find it in select field anymore.
if insertPlan.TableColNames.FindAstColName(colName.Name) {
continue
}
found := false
for _, field := range sel.Fields.Fields {
if colField, ok := field.Expr.(*ast.ColumnNameExpr); ok &&
(colName.Name.Schema.L == "" || colField.Name.Schema.L == "" || colName.Name.Schema.L == colField.Name.Schema.L) &&
(colName.Name.Table.L == "" || colField.Name.Table.L == "" || colName.Name.Table.L == colField.Name.Table.L) &&
colName.Name.Name.L == colField.Name.Name.L {
found = true
break
}
}
if found {
continue
}
sel.Fields.Fields = append(sel.Fields.Fields, &ast.SelectField{Expr: colName, Offset: len(sel.Fields.Fields)})
}
defer func(originSelFieldLen int) {
// Revert the change for ast. Because when we use the 'prepare' and 'execute' statement it will both build plan which will cause problem.
// You can see the issue #37187 for more details.
sel.Fields.Fields = sel.Fields.Fields[:originSelFieldLen]
}(actualColLen)
}
}
}
nodeW := resolve.NewNodeWWithCtx(insert.Select, b.resolveCtx)
selectPlan, err := b.Build(ctx, nodeW)
if err != nil {
return err
}
// Check to guarantee that the length of the row returned by select is equal to that of affectedValuesCols.
if (actualColLen == -1 && selectPlan.Schema().Len() != len(affectedValuesCols)) || (actualColLen != -1 && actualColLen != len(affectedValuesCols)) {
return plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(1)
}
// Check to guarantee that there's no generated column.
// This check should be done after the above one to make its behavior compatible with MySQL.
// For example, table t has two columns, namely a and b, and b is a generated column.
// "insert into t (b) select * from t" will raise an error that the column count is not matched.
// "insert into t select * from t" will raise an error that there's a generated column in the column list.
// If we do this check before the above one, "insert into t (b) select * from t" will raise an error
// that there's a generated column in the column list.
for _, col := range affectedValuesCols {
if col.IsGenerated() {
return plannererrors.ErrBadGeneratedColumn.GenWithStackByArgs(col.Name.O, insertPlan.Table.Meta().Name.O)
}
}
names := selectPlan.OutputNames()
insertPlan.SelectPlan, _, err = DoOptimize(ctx, b.ctx, b.optFlag, selectPlan.(base.LogicalPlan))
if err != nil {
return err
}
if actualColLen == -1 {
actualColLen = selectPlan.Schema().Len()
}
insertPlan.RowLen = actualColLen
// schema4NewRow is the schema for the newly created data record based on
// the result of the select statement.
schema4NewRow := expression.NewSchema(make([]*expression.Column, len(insertPlan.Table.Cols()))...)
names4NewRow := make(types.NameSlice, len(insertPlan.Table.Cols()))
// TODO: don't clone it.
for i := range actualColLen {
selCol := insertPlan.SelectPlan.Schema().Columns[i]
ordinal := affectedValuesCols[i].Offset
schema4NewRow.Columns[ordinal] = &expression.Column{}
*schema4NewRow.Columns[ordinal] = *selCol
schema4NewRow.Columns[ordinal].RetType = &types.FieldType{}
*schema4NewRow.Columns[ordinal].RetType = affectedValuesCols[i].FieldType
names4NewRow[ordinal] = names[i]
}
for i := range schema4NewRow.Columns {
if schema4NewRow.Columns[i] == nil {
schema4NewRow.Columns[i] = &expression.Column{UniqueID: insertPlan.SCtx().GetSessionVars().AllocPlanColumnID()}
names4NewRow[i] = types.EmptyName
}
}
insertPlan.Schema4OnDuplicate = expression.NewSchema(insertPlan.TableSchema.Columns...)
insertPlan.Schema4OnDuplicate.Append(insertPlan.SelectPlan.Schema().Columns[actualColLen:]...)
insertPlan.Schema4OnDuplicate.Append(schema4NewRow.Columns...)
insertPlan.Names4OnDuplicate = append(insertPlan.TableColNames.Shallow(), names[actualColLen:]...)
insertPlan.Names4OnDuplicate = append(insertPlan.Names4OnDuplicate, names4NewRow...)
return nil
}
func (b *PlanBuilder) buildLoadData(ctx context.Context, ld *ast.LoadDataStmt) (base.Plan, error) {
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
var (
err error
options = make([]*LoadDataOpt, 0, len(ld.Options))
)
for _, opt := range ld.Options {
loadDataOpt := LoadDataOpt{Name: opt.Name}
if opt.Value != nil {
loadDataOpt.Value, _, err = b.rewrite(ctx, opt.Value, mockTablePlan, nil, true)
if err != nil {
return nil, err
}
}
options = append(options, &loadDataOpt)
}
tnW := b.resolveCtx.GetTableName(ld.Table)
p := LoadData{
FileLocRef: ld.FileLocRef,
OnDuplicate: ld.OnDuplicate,
Path: ld.Path,
Format: ld.Format,
Table: tnW,
Charset: ld.Charset,
Columns: ld.Columns,
FieldsInfo: ld.FieldsInfo,
LinesInfo: ld.LinesInfo,
IgnoreLines: ld.IgnoreLines,
ColumnAssignments: ld.ColumnAssignments,
ColumnsAndUserVars: ld.ColumnsAndUserVars,
Options: options,
}.Init(b.ctx)
user := b.ctx.GetSessionVars().User
var insertErr, deleteErr error
if user != nil {
insertErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INSERT", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
deleteErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DELETE", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, p.Table.Schema.L, p.Table.Name.L, "", insertErr)
if p.OnDuplicate == ast.OnDuplicateKeyHandlingReplace {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DeletePriv, p.Table.Schema.L, p.Table.Name.L, "", deleteErr)
}
tableInfo := p.Table.TableInfo
tableInPlan, ok := b.is.TableByID(ctx, tableInfo.ID)
if !ok {
db := b.ctx.GetSessionVars().CurrentDB
return nil, infoschema.ErrTableNotExists.FastGenByArgs(db, tableInfo.Name.O)
}
schema, names, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), ast.NewCIStr(""), tableInfo)
if err != nil {
return nil, err
}
mockTablePlan.SetSchema(schema)
mockTablePlan.SetOutputNames(names)
p.GenCols, err = b.resolveGeneratedColumns(ctx, tableInPlan.Cols(), nil, mockTablePlan)
return p, err
}
var (
importIntoSchemaNames = []string{
"Job_ID", "Group_Key", "Data_Source", "Target_Table", "Table_ID",
"Phase", "Status", "Source_File_Size", "Imported_Rows",
"Result_Message", "Create_Time", "Start_Time", "End_Time", "Created_By", "Last_Update_Time",
"Cur_Step", "Cur_Step_Processed_Size", "Cur_Step_Total_Size", "Cur_Step_Progress_Pct", "Cur_Step_Speed", "Cur_Step_ETA",
}
// ImportIntoSchemaFTypes store the field types of the show import jobs schema.
ImportIntoSchemaFTypes = []byte{
mysql.TypeLonglong, mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeLonglong,
mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeLonglong,
mysql.TypeString, mysql.TypeTimestamp, mysql.TypeTimestamp, mysql.TypeTimestamp, mysql.TypeString, mysql.TypeTimestamp,
mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeString,
}
// ImportIntoFieldMap store the mapping from field names to their indices.
// As there are many test cases that use the index to check the result from
// `SHOW IMPORT JOBS`, this structure is used to avoid hardcoding these indexs,
// so adding new fields does not require modifying all the tests.
ImportIntoFieldMap = make(map[string]int)
showImportGroupsNames = []string{"Group_Key", "Total_Jobs",
"Pending", "Running", "Completed", "Failed", "Cancelled",
"First_Job_Create_Time", "Last_Job_Update_Time"}
showImportGroupsFTypes = []byte{mysql.TypeString, mysql.TypeLonglong,
mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong,
mysql.TypeTimestamp, mysql.TypeTimestamp}
// ImportIntoDataSource used inplannererrors.ErrLoadDataInvalidURI.
ImportIntoDataSource = "data source"
)
func init() {
for idx, name := range importIntoSchemaNames {
normalized := strings.ReplaceAll(name, "_", "")
ImportIntoFieldMap[normalized] = idx
}
}
var (
distributionJobsSchemaNames = []string{"Job_ID", "Database", "Table", "Partition_List", "Engine", "Rule", "Status",
"Timeout", "Create_Time", "Start_Time", "Finish_Time"}
distributionJobsSchedulerFTypes = []byte{mysql.TypeLonglong, mysql.TypeString, mysql.TypeString, mysql.TypeString,
mysql.TypeString, mysql.TypeString, mysql.TypeString, mysql.TypeString,
mysql.TypeTimestamp, mysql.TypeTimestamp, mysql.TypeTimestamp}
)
// importIntoCollAssignmentChecker implements ast.Visitor interface.
// It is used to check the column assignment expressions in IMPORT INTO statement.
// Currently, the import into column assignment only supports some simple expressions.
type importIntoCollAssignmentChecker struct {
idx int
err error
neededVars map[string]int
}
func newImportIntoCollAssignmentChecker() *importIntoCollAssignmentChecker {
return &importIntoCollAssignmentChecker{neededVars: make(map[string]int)}
}
// checkImportIntoColAssignments checks the column assignment expressions in IMPORT INTO statement.
func checkImportIntoColAssignments(assignments []*ast.Assignment) (map[string]int, error) {
checker := newImportIntoCollAssignmentChecker()
for i, assign := range assignments {
checker.idx = i
assign.Expr.Accept(checker)
if checker.err != nil {
break
}
}
return checker.neededVars, checker.err
}
// Enter implements ast.Visitor interface.
func (*importIntoCollAssignmentChecker) Enter(node ast.Node) (ast.Node, bool) {
return node, false
}
// Leave implements ast.Visitor interface.
func (v *importIntoCollAssignmentChecker) Leave(node ast.Node) (ast.Node, bool) {
switch n := node.(type) {
case *ast.ColumnNameExpr:
v.err = errors.Errorf("COLUMN reference is not supported in IMPORT INTO column assignment, index %d", v.idx)
return n, false
case *ast.SubqueryExpr:
v.err = errors.Errorf("subquery is not supported in IMPORT INTO column assignment, index %d", v.idx)
return n, false
case *ast.VariableExpr:
if n.IsSystem {
v.err = errors.Errorf("system variable is not supported in IMPORT INTO column assignment, index %d", v.idx)
return n, false
}
if n.Value != nil {
v.err = errors.Errorf("setting a variable in IMPORT INTO column assignment is not supported, index %d", v.idx)
return n, false
}
v.neededVars[strings.ToLower(n.Name)] = v.idx
case *ast.DefaultExpr:
v.err = errors.Errorf("FUNCTION default is not supported in IMPORT INTO column assignment, index %d", v.idx)
return n, false
case *ast.WindowFuncExpr:
v.err = errors.Errorf("window FUNCTION %s is not supported in IMPORT INTO column assignment, index %d", n.Name, v.idx)
return n, false
case *ast.AggregateFuncExpr:
v.err = errors.Errorf("aggregate FUNCTION %s is not supported in IMPORT INTO column assignment, index %d", n.F, v.idx)
return n, false
case *ast.FuncCallExpr:
fnName := n.FnName.L
switch fnName {
case ast.Grouping:
v.err = errors.Errorf("FUNCTION %s is not supported in IMPORT INTO column assignment, index %d", n.FnName.O, v.idx)
return n, false
case ast.GetVar:
if len(n.Args) > 0 {
val, ok := n.Args[0].(*driver.ValueExpr)
if !ok || val.Kind() != types.KindString {
v.err = errors.Errorf("the argument of getvar should be a constant string in IMPORT INTO column assignment, index %d", v.idx)
return n, false
}
v.neededVars[strings.ToLower(val.GetString())] = v.idx
}
default:
if !expression.IsFunctionSupported(fnName) {
v.err = errors.Errorf("FUNCTION %s is not supported in IMPORT INTO column assignment, index %d", n.FnName.O, v.idx)
return n, false
}
}
case *ast.ValuesExpr:
v.err = errors.Errorf("VALUES is not supported in IMPORT INTO column assignment, index %d", v.idx)
return n, false
}
return node, v.err == nil
}
func (b *PlanBuilder) buildImportInto(ctx context.Context, ld *ast.ImportIntoStmt) (base.Plan, error) {
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
var (
err error
options = make([]*LoadDataOpt, 0, len(ld.Options))
importFromServer bool
)
if ld.Select == nil {
u, err := url.Parse(ld.Path)
if err != nil {
return nil, exeerrors.ErrLoadDataInvalidURI.FastGenByArgs(ImportIntoDataSource, err.Error())
}
importFromServer = objstore.IsLocal(u)
// for SEM v2, they are checked by configured rules.
if semv1.IsEnabled() {
if importFromServer {
return nil, plannererrors.ErrNotSupportedWithSem.GenWithStackByArgs("IMPORT INTO from server disk")
}
if kerneltype.IsNextGen() && objstore.IsS3(u) {
if err := checkNextGenS3PathWithSem(u); err != nil {
return nil, err
}
}
}
// a nextgen cluster might be shared by multiple tenants, and they might
// share the same AWS role to access import-into source data bucket, this
// external ID can be used to restrict the access only to the current tenant.
// when SEM enabled, we need set it.
if kerneltype.IsNextGen() && sem.IsEnabled() && objstore.IsS3(u) {
values := u.Query()
values.Set(s3like.S3ExternalID, config.GetGlobalKeyspaceName())
u.RawQuery = values.Encode()
ld.Path = u.String()
}
}
for _, opt := range ld.Options {
loadDataOpt := LoadDataOpt{Name: opt.Name}
if opt.Value != nil {
loadDataOpt.Value, _, err = b.rewrite(ctx, opt.Value, mockTablePlan, nil, true)
if err != nil {
return nil, err
}
}
options = append(options, &loadDataOpt)
}
neededVars, err := checkImportIntoColAssignments(ld.ColumnAssignments)
if err != nil {
return nil, err
}
for _, v := range ld.ColumnsAndUserVars {
userVar := v.UserVar
if userVar == nil {
continue
}
delete(neededVars, strings.ToLower(userVar.Name))
}
if len(neededVars) > 0 {
valuesStr := make([]string, 0, len(neededVars))
for _, v := range neededVars {
valuesStr = append(valuesStr, strconv.Itoa(v))
}
return nil, errors.Errorf(
"column assignment cannot use variables set outside IMPORT INTO statement, index %s",
strings.Join(valuesStr, ","),
)
}
tnW := b.resolveCtx.GetTableName(ld.Table)
if tnW.TableInfo.TempTableType != model.TempTableNone {
return nil, errors.Errorf("IMPORT INTO does not support temporary table")
} else if tnW.TableInfo.TableCacheStatusType != model.TableCacheStatusDisable {
return nil, errors.Errorf("IMPORT INTO does not support cached table")
}
p := ImportInto{
Path: ld.Path,
Format: ld.Format,
Table: tnW,
ColumnAssignments: ld.ColumnAssignments,
ColumnsAndUserVars: ld.ColumnsAndUserVars,
Options: options,
Stmt: ld.Text(),
}.Init(b.ctx)
user := b.ctx.GetSessionVars().User
// IMPORT INTO need full DML privilege of the target table
// to support add-index by SQL, we need ALTER privilege
var selectErr, updateErr, insertErr, deleteErr, alterErr error
if user != nil {
selectErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SELECT", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
updateErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("UPDATE", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
insertErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INSERT", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
deleteErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DELETE", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
alterErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ALTER", user.AuthUsername, user.AuthHostname, p.Table.Name.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, p.Table.Schema.L, p.Table.Name.L, "", selectErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.UpdatePriv, p.Table.Schema.L, p.Table.Name.L, "", updateErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, p.Table.Schema.L, p.Table.Name.L, "", insertErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DeletePriv, p.Table.Schema.L, p.Table.Name.L, "", deleteErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, p.Table.Schema.L, p.Table.Name.L, "", alterErr)
if importFromServer {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.FilePriv, "", "", "", plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("FILE"))
}
tableInfo := p.Table.TableInfo
// we use the latest IS to support IMPORT INTO dst FROM SELECT * FROM src AS OF TIMESTAMP '2020-01-01 00:00:00'
// Note: we need to get p.Table when preprocessing, at that time, IS of session
// transaction is used, if the session ctx is already in snapshot read using tidb_snapshot, we might
// not get the schema or get a stale schema of the target table, so we don't
// support set 'tidb_snapshot' first and then import into the target table.
//
// tidb_read_staleness can be used to do stale read too, it's allowed as long as
// TableInfo.ID matches with the latest schema.
latestIS := b.ctx.GetLatestInfoSchema().(infoschema.InfoSchema)
tableInPlan, ok := latestIS.TableByID(ctx, tableInfo.ID)
if !ok {
// adaptor.handleNoDelayExecutor has a similar check, but we want to give
// a more specific error message here.
if b.ctx.GetSessionVars().SnapshotTS != 0 {
return nil, errors.New("can not execute IMPORT statement when 'tidb_snapshot' is set")
}
db := b.ctx.GetSessionVars().CurrentDB
return nil, infoschema.ErrTableNotExists.FastGenByArgs(db, tableInfo.Name.O)
}
schema, names, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), ast.NewCIStr(""), tableInfo)
if err != nil {
return nil, err
}
mockTablePlan.SetSchema(schema)
mockTablePlan.SetOutputNames(names)
p.GenCols, err = b.resolveGeneratedColumns(ctx, tableInPlan.Cols(), nil, mockTablePlan)
if err != nil {
return nil, err
}
if ld.Select != nil {
// privilege of tables in select will be checked here
nodeW := resolve.NewNodeWWithCtx(ld.Select, b.resolveCtx)
selectPlan, err2 := b.Build(ctx, nodeW)
if err2 != nil {
return nil, err2
}
// it's allowed to use IMPORT INTO t FROM SELECT * FROM t
// as we pre-check that the target table must be empty.
if (len(ld.ColumnsAndUserVars) > 0 && len(selectPlan.Schema().Columns) != len(ld.ColumnsAndUserVars)) ||
(len(ld.ColumnsAndUserVars) == 0 && len(selectPlan.Schema().Columns) != len(tableInPlan.VisibleCols())) {
return nil, plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(1)
}
p.SelectPlan, _, err2 = DoOptimize(ctx, b.ctx, b.optFlag, selectPlan.(base.LogicalPlan))
if err2 != nil {
return nil, err2
}
} else {
outputSchema, outputFields := convert2OutputSchemasAndNames(importIntoSchemaNames, ImportIntoSchemaFTypes, []uint{})
p.SetSchemaAndNames(outputSchema, outputFields)
}
return p, nil
}
func (*PlanBuilder) buildLoadStats(ld *ast.LoadStatsStmt) base.Plan {
p := &LoadStats{Path: ld.Path}
return p
}
func (b *PlanBuilder) buildRefreshStats(rs *ast.RefreshStatsStmt) (base.Plan, error) {
if err := fillDefaultDBForRefreshStats(b.ctx, rs); err != nil {
return nil, err
}
rs.Dedup()
b.requireSelectOrRestoreAdminPrivForRefreshStats(rs)
p := &Simple{
Statement: rs,
IsFromRemote: false,
ResolveCtx: b.resolveCtx,
}
return p, nil
}
func fillDefaultDBForRefreshStats(ctx base.PlanContext, rs *ast.RefreshStatsStmt) error {
if len(rs.RefreshObjects) == 0 {
return nil
}
currentDB := ctx.GetSessionVars().CurrentDB
for _, obj := range rs.RefreshObjects {
if obj.RefreshObjectScope != ast.RefreshObjectScopeTable {
continue
}
if obj.DBName.L != "" {
continue
}
if currentDB == "" {
return plannererrors.ErrNoDB
}
obj.DBName = ast.NewCIStr(currentDB)
}
return nil
}
func (b *PlanBuilder) requireSelectOrRestoreAdminPrivForRefreshStats(rs *ast.RefreshStatsStmt) {
if len(rs.RefreshObjects) == 0 {
intest.Assert(len(rs.RefreshObjects) > 0, "RefreshStatsStmt.RefreshObjects should not be empty")
return
}
checker := privilege.GetPrivilegeManager(b.ctx)
if checker != nil {
activeRoles := b.ctx.GetSessionVars().ActiveRoles
if checker.RequestDynamicVerification(activeRoles, "RESTORE_ADMIN", false) {
return
}
}
user := b.ctx.GetSessionVars().User
for _, obj := range rs.RefreshObjects {
switch obj.RefreshObjectScope {
case ast.RefreshObjectScopeGlobal:
var err error
if user != nil {
err = plannererrors.ErrPrivilegeCheckFail.GenWithStackByArgs("SELECT")
} else {
err = plannererrors.ErrPrivilegeCheckFail
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, "", "", "", err)
return
case ast.RefreshObjectScopeDatabase:
var err error
if user != nil {
err = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(user.AuthUsername, user.AuthHostname, obj.DBName.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, obj.DBName.L, "", "", err)
case ast.RefreshObjectScopeTable:
dbName := obj.DBName.L
var err error
if user != nil {
err = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SELECT", user.AuthUsername, user.AuthHostname, obj.TableName.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, dbName, obj.TableName.L, "", err)
}
}
}
// buildLockStats requires INSERT and SELECT privilege for the tables same as buildAnalyze.
func (b *PlanBuilder) buildLockStats(ld *ast.LockStatsStmt) base.Plan {
p := &LockStats{
Tables: ld.Tables,
}
b.requireInsertAndSelectPriv(ld.Tables)
return p
}
// buildUnlockStats requires INSERT and SELECT privilege for the tables same as buildAnalyze.
func (b *PlanBuilder) buildUnlockStats(ld *ast.UnlockStatsStmt) base.Plan {
p := &UnlockStats{
Tables: ld.Tables,
}
b.requireInsertAndSelectPriv(ld.Tables)
return p
}
// requireInsertAndSelectPriv requires INSERT and SELECT privilege for the tables.
func (b *PlanBuilder) requireInsertAndSelectPriv(tables []*ast.TableName) {
for _, tbl := range tables {
user := b.ctx.GetSessionVars().User
var insertErr, selectErr error
if user != nil {
insertErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INSERT", user.AuthUsername, user.AuthHostname, tbl.Name.O)
selectErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("SELECT", user.AuthUsername, user.AuthHostname, tbl.Name.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, tbl.Schema.L, tbl.Name.L, "", insertErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, tbl.Schema.L, tbl.Name.L, "", selectErr)
}
}
var ruleList = []string{"leader-scatter", "peer-scatter", "learner-scatter"}
var engineList = []string{"tikv", "tiflash"}
func (b *PlanBuilder) buildDistributeTable(node *ast.DistributeTableStmt) (base.Plan, error) {
tnW := b.resolveCtx.GetTableName(node.Table)
tblInfo := tnW.TableInfo
if !slices.Contains(ruleList, node.Rule) {
return nil, plannererrors.ErrWrongArguments.GenWithStackByArgs("rule must be leader-scatter, peer-scatter or learner-scatter")
}
if !slices.Contains(engineList, node.Engine) {
return nil, plannererrors.ErrWrongArguments.GenWithStackByArgs("engine must be tikv or tiflash")
}
if node.Engine == "tiflash" && node.Rule != "learner-scatter" {
return nil, plannererrors.ErrWrongArguments.GenWithStackByArgs("the rule of tiflash must be learner-scatter")
}
plan := &DistributeTable{
TableInfo: tblInfo,
PartitionNames: node.PartitionNames,
Engine: node.Engine,
Rule: node.Rule,
Timeout: node.Timeout,
}
plan.SetSchemaAndNames(buildDistributeTableSchema())
return plan, nil
}
func (b *PlanBuilder) buildSplitRegion(node *ast.SplitRegionStmt) (base.Plan, error) {
tnW := b.resolveCtx.GetTableName(node.Table)
if tnW.TableInfo.TempTableType != model.TempTableNone {
return nil, plannererrors.ErrOptOnTemporaryTable.GenWithStackByArgs("split table")
}
if node.SplitSyntaxOpt != nil && node.SplitSyntaxOpt.HasPartition && tnW.TableInfo.Partition == nil {
return nil, plannererrors.ErrPartitionClauseOnNonpartitioned
}
if len(node.IndexName.L) != 0 {
return b.buildSplitIndexRegion(node)
}
return b.buildSplitTableRegion(node)
}
func (b *PlanBuilder) buildSplitIndexRegion(node *ast.SplitRegionStmt) (base.Plan, error) {
tnW := b.resolveCtx.GetTableName(node.Table)
tblInfo := tnW.TableInfo
if node.IndexName.L == strings.ToLower(mysql.PrimaryKeyName) &&
(tblInfo.IsCommonHandle || tblInfo.PKIsHandle) {
return nil, plannererrors.ErrKeyDoesNotExist.FastGen("unable to split clustered index, please split table instead.")
}
indexInfo := tblInfo.FindIndexByName(node.IndexName.L)
if indexInfo == nil {
return nil, plannererrors.ErrKeyDoesNotExist.GenWithStackByArgs(node.IndexName, tblInfo.Name)
}
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
schema, names, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), node.Table.Schema, tblInfo)
if err != nil {
return nil, err
}
mockTablePlan.SetSchema(schema)
mockTablePlan.SetOutputNames(names)
p := &SplitRegion{
TableInfo: tblInfo,
PartitionNames: node.PartitionNames,
IndexInfo: indexInfo,
}
p.SetOutputNames(names)
p.SetSchemaAndNames(buildSplitRegionsSchema())
// Split index regions by user specified value lists.
if len(node.SplitOpt.ValueLists) > 0 {
indexValues := make([][]types.Datum, 0, len(node.SplitOpt.ValueLists))
for i, valuesItem := range node.SplitOpt.ValueLists {
if len(valuesItem) > len(indexInfo.Columns) {
return nil, plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(i + 1)
}
values, err := b.convertValue2ColumnType(valuesItem, mockTablePlan, indexInfo, tblInfo)
if err != nil {
return nil, err
}
indexValues = append(indexValues, values)
}
p.ValueLists = indexValues
return p, nil
}
// Split index regions by lower, upper value.
checkLowerUpperValue := func(valuesItem []ast.ExprNode, name string) ([]types.Datum, error) {
if len(valuesItem) == 0 {
return nil, errors.Errorf("Split index `%v` region %s value count should more than 0", indexInfo.Name, name)
}
if len(valuesItem) > len(indexInfo.Columns) {
return nil, errors.Errorf("Split index `%v` region column count doesn't match value count at %v", indexInfo.Name, name)
}
return b.convertValue2ColumnType(valuesItem, mockTablePlan, indexInfo, tblInfo)
}
lowerValues, err := checkLowerUpperValue(node.SplitOpt.Lower, "lower")
if err != nil {
return nil, err
}
upperValues, err := checkLowerUpperValue(node.SplitOpt.Upper, "upper")
if err != nil {
return nil, err
}
p.Lower = lowerValues
p.Upper = upperValues
maxSplitRegionNum := int64(config.GetGlobalConfig().SplitRegionMaxNum)
if node.SplitOpt.Num > maxSplitRegionNum {
return nil, errors.Errorf("Split index region num exceeded the limit %v", maxSplitRegionNum)
} else if node.SplitOpt.Num < 1 {
return nil, errors.Errorf("Split index region num should more than 0")
}
p.Num = int(node.SplitOpt.Num)
return p, nil
}
func (b *PlanBuilder) convertValue2ColumnType(valuesItem []ast.ExprNode, mockTablePlan base.LogicalPlan, indexInfo *model.IndexInfo, tblInfo *model.TableInfo) ([]types.Datum, error) {
values := make([]types.Datum, 0, len(valuesItem))
for j, valueItem := range valuesItem {
colOffset := indexInfo.Columns[j].Offset
value, err := b.convertValue(valueItem, mockTablePlan, tblInfo.Columns[colOffset])
if err != nil {
return nil, err
}
values = append(values, value)
}
return values, nil
}
func (b *PlanBuilder) convertValue(valueItem ast.ExprNode, mockTablePlan base.LogicalPlan, col *model.ColumnInfo) (d types.Datum, err error) {
var expr expression.Expression
switch x := valueItem.(type) {
case *driver.ValueExpr:
expr = &expression.Constant{
Value: x.Datum,
RetType: &x.Type,
}
default:
expr, _, err = b.rewrite(context.TODO(), valueItem, mockTablePlan, nil, true)
if err != nil {
return d, err
}
}
constant, ok := expr.(*expression.Constant)
if !ok {
return d, errors.New("Expect constant values")
}
value, err := constant.Eval(b.ctx.GetExprCtx().GetEvalCtx(), chunk.Row{})
if err != nil {
return d, err
}
d, err = value.ConvertTo(b.ctx.GetSessionVars().StmtCtx.TypeCtx(), &col.FieldType)
if err != nil {
if !types.ErrTruncated.Equal(err) && !types.ErrTruncatedWrongVal.Equal(err) && !types.ErrBadNumber.Equal(err) {
return d, err
}
valStr, err1 := value.ToString()
if err1 != nil {
return d, err
}
return d, types.ErrTruncated.GenWithStack("Incorrect value: '%-.128s' for column '%.192s'", valStr, col.Name.O)
}
return d, nil
}
func (b *PlanBuilder) buildSplitTableRegion(node *ast.SplitRegionStmt) (base.Plan, error) {
tnW := b.resolveCtx.GetTableName(node.Table)
tblInfo := tnW.TableInfo
handleColInfos := buildHandleColumnInfos(tblInfo)
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
schema, names, err := expression.TableInfo2SchemaAndNames(b.ctx.GetExprCtx(), node.Table.Schema, tblInfo)
if err != nil {
return nil, err
}
mockTablePlan.SetSchema(schema)
mockTablePlan.SetOutputNames(names)
p := &SplitRegion{
TableInfo: tblInfo,
PartitionNames: node.PartitionNames,
}
p.SetSchemaAndNames(buildSplitRegionsSchema())
if len(node.SplitOpt.ValueLists) > 0 {
values := make([][]types.Datum, 0, len(node.SplitOpt.ValueLists))
for i, valuesItem := range node.SplitOpt.ValueLists {
data, err := convertValueListToData(valuesItem, handleColInfos, i, b, mockTablePlan)
if err != nil {
return nil, err
}
values = append(values, data)
}
p.ValueLists = values
return p, nil
}
p.Lower, err = convertValueListToData(node.SplitOpt.Lower, handleColInfos, lowerBound, b, mockTablePlan)
if err != nil {
return nil, err
}
p.Upper, err = convertValueListToData(node.SplitOpt.Upper, handleColInfos, upperBound, b, mockTablePlan)
if err != nil {
return nil, err
}
maxSplitRegionNum := int64(config.GetGlobalConfig().SplitRegionMaxNum)
if node.SplitOpt.Num > maxSplitRegionNum {
return nil, errors.Errorf("Split table region num exceeded the limit %v", maxSplitRegionNum)
} else if node.SplitOpt.Num < 1 {
return nil, errors.Errorf("Split table region num should more than 0")
}
p.Num = int(node.SplitOpt.Num)
return p, nil
}
func buildHandleColumnInfos(tblInfo *model.TableInfo) []*model.ColumnInfo {
switch {
case tblInfo.PKIsHandle:
if col := tblInfo.GetPkColInfo(); col != nil {
return []*model.ColumnInfo{col}
}
case tblInfo.IsCommonHandle:
pkIdx := tables.FindPrimaryIndex(tblInfo)
pkCols := make([]*model.ColumnInfo, 0, len(pkIdx.Columns))
cols := tblInfo.Columns
for _, idxCol := range pkIdx.Columns {
pkCols = append(pkCols, cols[idxCol.Offset])
}
return pkCols
default:
return []*model.ColumnInfo{model.NewExtraHandleColInfo()}
}
return nil
}
const (
lowerBound int = -1
upperBound int = -2
)
func convertValueListToData(valueList []ast.ExprNode, handleColInfos []*model.ColumnInfo, rowIdx int,
b *PlanBuilder, mockTablePlan *logicalop.LogicalTableDual) ([]types.Datum, error) {
if len(valueList) != len(handleColInfos) {
var err error
switch rowIdx {
case lowerBound:
err = errors.Errorf("Split table region lower value count should be %d", len(handleColInfos))
case upperBound:
err = errors.Errorf("Split table region upper value count should be %d", len(handleColInfos))
default:
err = plannererrors.ErrWrongValueCountOnRow.GenWithStackByArgs(rowIdx)
}
return nil, err
}
data := make([]types.Datum, 0, len(handleColInfos))
for i, v := range valueList {
convertedDatum, err := b.convertValue(v, mockTablePlan, handleColInfos[i])
if err != nil {
return nil, err
}
if convertedDatum.IsNull() {
return nil, plannererrors.ErrBadNull.GenWithStackByArgs(handleColInfos[i].Name.O)
}
data = append(data, convertedDatum)
}
return data, nil
}
type userVariableChecker struct {
hasUserVariables bool
}
func (e *userVariableChecker) Enter(in ast.Node) (ast.Node, bool) {
if _, ok := in.(*ast.VariableExpr); ok {
e.hasUserVariables = true
return in, true
}
return in, false
}
func (*userVariableChecker) Leave(in ast.Node) (ast.Node, bool) {
return in, true
}
// Check for UserVariables
func checkForUserVariables(in ast.Node) error {
v := &userVariableChecker{hasUserVariables: false}
_, ok := in.Accept(v)
if !ok || v.hasUserVariables {
return dbterror.ErrViewSelectVariable
}
return nil
}
func (b *PlanBuilder) buildDDL(ctx context.Context, node ast.DDLNode) (base.Plan, error) {
var authErr error
switch v := node.(type) {
case *ast.AlterDatabaseStmt:
if v.AlterDefaultDatabase {
v.Name = ast.NewCIStr(b.ctx.GetSessionVars().CurrentDB)
}
if v.Name.O == "" {
return nil, plannererrors.ErrNoDB
}
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Name.O)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, v.Name.L, "", "", authErr)
case *ast.AlterTableStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ALTER", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
dbName := getLowerDB(v.Table.Schema, b.ctx.GetSessionVars())
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, dbName,
v.Table.Name.L, "", authErr)
for _, spec := range v.Specs {
if spec.Tp == ast.AlterTableRenameTable || spec.Tp == ast.AlterTableExchangePartition {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, dbName,
v.Table.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, spec.NewTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, dbName,
spec.NewTable.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INSERT", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, spec.NewTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, dbName,
spec.NewTable.Name.L, "", authErr)
} else if spec.Tp == ast.AlterTableDropPartition || spec.Tp == ast.AlterTableTruncatePartition {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.Table.Schema.L,
v.Table.Name.L, "", authErr)
} else if spec.Tp == ast.AlterTableWriteable {
b.visitInfo[0].alterWritable = true
} else if spec.Tp == ast.AlterTableAddStatistics {
var selectErr, insertErr error
user := b.ctx.GetSessionVars().User
if user != nil {
selectErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ADD STATS_EXTENDED", user.AuthUsername,
user.AuthHostname, v.Table.Name.L)
insertErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ADD STATS_EXTENDED", user.AuthUsername,
user.AuthHostname, "stats_extended")
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, v.Table.Schema.L,
v.Table.Name.L, "", selectErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, mysql.SystemDB,
"stats_extended", "", insertErr)
} else if spec.Tp == ast.AlterTableDropStatistics {
user := b.ctx.GetSessionVars().User
if user != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP STATS_EXTENDED", user.AuthUsername,
user.AuthHostname, "stats_extended")
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.UpdatePriv, mysql.SystemDB,
"stats_extended", "", authErr)
} else if spec.Tp == ast.AlterTableAddConstraint {
if b.ctx.GetSessionVars().User != nil && spec.Constraint != nil &&
spec.Constraint.Tp == ast.ConstraintForeignKey && spec.Constraint.Refer != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("REFERENCES", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, spec.Constraint.Refer.Table.Name.L)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ReferencesPriv, spec.Constraint.Refer.Table.Schema.L,
spec.Constraint.Refer.Table.Name.L, "", authErr)
}
}
}
case *ast.AlterSequenceStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ALTER", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Name.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, v.Name.Schema.L,
v.Name.Name.L, "", authErr)
case *ast.CreateDatabaseStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Name)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.Name.L,
"", "", authErr)
case *ast.CreateIndexStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INDEX", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.IndexPriv, v.Table.Schema.L,
v.Table.Name.L, "", authErr)
case *ast.CreateTableStmt:
if v.TemporaryKeyword != ast.TemporaryNone {
for _, cons := range v.Constraints {
if cons.Tp == ast.ConstraintForeignKey {
return nil, infoschema.ErrCannotAddForeign
}
}
}
if b.ctx.GetSessionVars().User != nil {
// This is tricky here: we always need the visitInfo because it's not only used in privilege checks, and we
// must pass the table name. However, the privilege check is towards the database. We'll deal with it later.
if v.TemporaryKeyword == ast.TemporaryLocal {
authErr = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Schema.L)
} else {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
for _, cons := range v.Constraints {
if cons.Tp == ast.ConstraintForeignKey && cons.Refer != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("REFERENCES", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, cons.Refer.Table.Name.L)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.ReferencesPriv, cons.Refer.Table.Schema.L,
cons.Refer.Table.Name.L, "", authErr)
}
}
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.Table.Schema.L,
v.Table.Name.L, "", authErr)
if v.ReferTable != nil {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.ReferTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, v.ReferTable.Schema.L,
v.ReferTable.Name.L, "", authErr)
}
case *ast.CreateViewStmt:
err := checkForUserVariables(v.Select)
if err != nil {
return nil, err
}
b.isCreateView = true
b.capFlag |= canExpandAST | renameView
b.renamingViewName = v.ViewName.Schema.L + "." + v.ViewName.Name.L
defer func() {
b.capFlag &= ^canExpandAST
b.capFlag &= ^renameView
b.isCreateView = false
}()
if stmt := findStmtAsViewSchema(v); stmt != nil {
stmt.AsViewSchema = true
}
nodeW := resolve.NewNodeWWithCtx(v.Select, b.resolveCtx)
plan, err := b.Build(ctx, nodeW)
if err != nil {
return nil, err
}
schema := plan.Schema()
names := plan.OutputNames()
if v.Cols == nil {
adjustOverlongViewColname(plan.(base.LogicalPlan))
v.Cols = make([]ast.CIStr, len(schema.Columns))
for i, name := range names {
v.Cols[i] = name.ColName
}
}
if len(v.Cols) != schema.Len() {
return nil, dbterror.ErrViewWrongList
}
var authCreateErr, authDropErr error
if user := b.ctx.GetSessionVars().User; user != nil {
authCreateErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE VIEW", user.AuthUsername,
user.AuthHostname, v.ViewName.Name.L)
if v.OrReplace {
authDropErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", user.AuthUsername,
user.AuthHostname, v.ViewName.Name.L)
}
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreateViewPriv, v.ViewName.Schema.L,
v.ViewName.Name.L, "", authCreateErr)
if v.OrReplace {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.ViewName.Schema.L,
v.ViewName.Name.L, "", authDropErr)
}
if v.Definer.CurrentUser && b.ctx.GetSessionVars().User != nil {
v.Definer = b.ctx.GetSessionVars().User
}
if b.ctx.GetSessionVars().User != nil && v.Definer.String() != b.ctx.GetSessionVars().User.String() {
err = plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER")
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "",
"", "", err)
}
case *ast.CreateSequenceStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Name.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.Name.Schema.L,
v.Name.Name.L, "", authErr)
case *ast.DropDatabaseStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Name)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.Name.L,
"", "", authErr)
case *ast.DropIndexStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INDEX", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.IndexPriv, v.Table.Schema.L,
v.Table.Name.L, "", authErr)
case *ast.DropTableStmt:
for _, tableVal := range v.Tables {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, tableVal.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, tableVal.Schema.L,
tableVal.Name.L, "", authErr)
}
case *ast.DropSequenceStmt:
for _, sequence := range v.Sequences {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, sequence.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, sequence.Schema.L,
sequence.Name.L, "", authErr)
}
case *ast.TruncateTableStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.Table.Schema.L,
v.Table.Name.L, "", authErr)
case *ast.RenameTableStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ALTER", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.TableToTables[0].OldTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, v.TableToTables[0].OldTable.Schema.L,
v.TableToTables[0].OldTable.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.TableToTables[0].OldTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.TableToTables[0].OldTable.Schema.L,
v.TableToTables[0].OldTable.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.TableToTables[0].NewTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.TableToTables[0].NewTable.Schema.L,
v.TableToTables[0].NewTable.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("INSERT", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.TableToTables[0].NewTable.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.InsertPriv, v.TableToTables[0].NewTable.Schema.L,
v.TableToTables[0].NewTable.Name.L, "", authErr)
case *ast.RecoverTableStmt:
if v.Table == nil {
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
} else {
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.Table.Schema.L, v.Table.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.Table.Schema.L, v.Table.Name.L, "", authErr)
}
case *ast.FlashBackTableStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("CREATE", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.Table.Schema.L, v.Table.Name.L, "", authErr)
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("DROP", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.Table.Schema.L, v.Table.Name.L, "", authErr)
case *ast.FlashBackDatabaseStmt:
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrDBaccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, v.DBName.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.CreatePriv, v.DBName.L, "", "", authErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.DropPriv, v.DBName.L, "", "", authErr)
case *ast.FlashBackToTimestampStmt:
// Flashback cluster can only be executed by user with `super` privilege.
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
case *ast.LockTablesStmt:
user := b.ctx.GetSessionVars().User
for _, lock := range v.TableLocks {
var lockAuthErr, selectAuthErr error
if user != nil {
lockAuthErr = plannererrors.ErrDBaccessDenied.FastGenByArgs(user.AuthUsername, user.AuthHostname, lock.Table.Schema.L)
selectAuthErr = plannererrors.ErrTableaccessDenied.FastGenByArgs("SELECT", user.AuthUsername, user.AuthHostname, lock.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.LockTablesPriv, lock.Table.Schema.L, lock.Table.Name.L, "", lockAuthErr)
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SelectPriv, lock.Table.Schema.L, lock.Table.Name.L, "", selectAuthErr)
}
case *ast.CleanupTableLockStmt:
// This command can only be executed by administrator.
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
case *ast.RepairTableStmt:
// Repair table command can only be executed by administrator.
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", nil)
case *ast.DropPlacementPolicyStmt, *ast.CreatePlacementPolicyStmt, *ast.AlterPlacementPolicyStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or PLACEMENT_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"PLACEMENT_ADMIN"}, false, err)
case *ast.CreateResourceGroupStmt, *ast.DropResourceGroupStmt, *ast.AlterResourceGroupStmt:
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("SUPER or RESOURCE_GROUP_ADMIN")
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, []string{"RESOURCE_GROUP_ADMIN"}, false, err)
case *ast.OptimizeTableStmt:
return nil, dbterror.ErrGeneralUnsupportedDDL.GenWithStack("OPTIMIZE TABLE is not supported")
}
p := &DDL{Statement: node}
return p, nil
}
const (
// TraceFormatRow indicates row tracing format.
TraceFormatRow = "row"
// TraceFormatJSON indicates json tracing format.
TraceFormatJSON = "json"
// TraceFormatLog indicates log tracing format.
TraceFormatLog = "log"
// TracePlanTargetEstimation indicates CE trace target for optimizer trace.
TracePlanTargetEstimation = "estimation"
// TracePlanTargetDebug indicates debug trace target for optimizer trace.
TracePlanTargetDebug = "debug"
)
// buildTrace builds a trace plan. Inside this method, it first optimize the
// underlying query and then constructs a schema, which will be used to constructs
// rows result.
func (b *PlanBuilder) buildTrace(trace *ast.TraceStmt) (base.Plan, error) {
p := &Trace{
StmtNode: trace.Stmt,
ResolveCtx: b.resolveCtx,
Format: trace.Format,
OptimizerTrace: trace.TracePlan,
OptimizerTraceTarget: trace.TracePlanTarget,
}
// TODO: forbid trace plan if the statement isn't select read-only statement
if trace.TracePlan {
switch trace.TracePlanTarget {
case "":
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "Dump_link", mysql.TypeVarchar, 128))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
case TracePlanTargetEstimation:
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "CE_trace", mysql.TypeVarchar, mysql.MaxBlobWidth))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
case TracePlanTargetDebug:
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "Debug_trace", mysql.TypeVarchar, mysql.MaxBlobWidth))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
default:
return nil, errors.New("trace plan target should only be 'estimation'")
}
return p, nil
}
switch trace.Format {
case TraceFormatRow:
schema := newColumnsWithNames(3)
schema.Append(buildColumnWithName("", "operation", mysql.TypeString, mysql.MaxBlobWidth))
schema.Append(buildColumnWithName("", "startTS", mysql.TypeString, mysql.MaxBlobWidth))
schema.Append(buildColumnWithName("", "duration", mysql.TypeString, mysql.MaxBlobWidth))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
case TraceFormatJSON:
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "operation", mysql.TypeString, mysql.MaxBlobWidth))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
case TraceFormatLog:
schema := newColumnsWithNames(4)
schema.Append(buildColumnWithName("", "time", mysql.TypeTimestamp, mysql.MaxBlobWidth))
schema.Append(buildColumnWithName("", "event", mysql.TypeString, mysql.MaxBlobWidth))
schema.Append(buildColumnWithName("", "tags", mysql.TypeString, mysql.MaxBlobWidth))
schema.Append(buildColumnWithName("", "spanName", mysql.TypeString, mysql.MaxBlobWidth))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
default:
return nil, errors.New("trace format should be one of 'row', 'log' or 'json'")
}
return p, nil
}
func (b *PlanBuilder) buildExplainPlan(targetPlan base.Plan, format string, explainBriefBinary string, analyze, explore bool, execStmt ast.StmtNode, runtimeStats *execdetails.RuntimeStatsColl, sqlDigest string) (base.Plan, error) {
format = strings.ToLower(format)
if format == types.ExplainFormatTrueCardCost && !analyze {
return nil, errors.Errorf("'explain format=%v' cannot work without 'analyze', please use 'explain analyze format=%v'", format, format)
}
p := &Explain{
TargetPlan: targetPlan,
Format: format,
Analyze: analyze,
Explore: explore,
SQLDigest: sqlDigest,
ExecStmt: execStmt,
BriefBinaryPlan: explainBriefBinary,
RuntimeStatsColl: runtimeStats,
}
p.SetSCtx(b.ctx)
return p, p.prepareSchema()
}
// buildExplainFor gets *last* (maybe running or finished) query plan from connection #connection id.
// See https://dev.mysql.com/doc/refman/8.0/en/explain-for-connection.html.
func (b *PlanBuilder) buildExplainFor(explainFor *ast.ExplainForStmt) (base.Plan, error) {
processInfo, ok := b.ctx.GetSessionManager().GetProcessInfo(explainFor.ConnectionID)
if !ok {
return nil, plannererrors.ErrNoSuchThread.GenWithStackByArgs(explainFor.ConnectionID)
}
if b.ctx.GetSessionVars() != nil && b.ctx.GetSessionVars().User != nil {
if b.ctx.GetSessionVars().User.Username != processInfo.User {
err := plannererrors.ErrAccessDenied.GenWithStackByArgs(b.ctx.GetSessionVars().User.Username, b.ctx.GetSessionVars().User.Hostname)
// Different from MySQL's behavior and document.
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.SuperPriv, "", "", "", err)
}
}
targetPlan, ok := processInfo.Plan.(base.Plan)
explainForFormat := strings.ToLower(explainFor.Format)
if !ok || targetPlan == nil {
return &Explain{Format: explainForFormat}, nil
}
// TODO: support other explain formats for connection
if explainForFormat != types.ExplainFormatBrief && explainForFormat != types.ExplainFormatROW && explainForFormat != types.ExplainFormatVerbose {
return nil, errors.Errorf("explain format '%s' for connection is not supported now", explainForFormat)
}
return b.buildExplainPlan(targetPlan, explainForFormat, processInfo.BriefBinaryPlan, false, false, nil, processInfo.RuntimeStatsColl, "")
}
// getHintedStmtThroughPlanDigest gets the hinted SQL like `select /*+ ... */ * from t where ...` from `stmt_summary`
// for `explain [analyze] <plan_digest>` statements.
func getHintedStmtThroughPlanDigest(ctx base.PlanContext, planDigest string) (stmt ast.StmtNode, err error) {
err = domain.GetDomain(ctx).AdvancedSysSessionPool().WithSession(func(se *syssession.Session) error {
return se.WithSessionContext(func(sctx sessionctx.Context) error {
defer func(warnings []stmtctx.SQLWarn) {
sctx.GetSessionVars().StmtCtx.SetWarnings(warnings)
}(sctx.GetSessionVars().StmtCtx.GetWarnings())
// The warnings will be broken in fetchRecordFromClusterStmtSummary(), so we need to save and restore it to make the
// warnings for repeated SQL Digest work.
schema, query, planHint, characterSet, collation, err := fetchRecordFromClusterStmtSummary(sctx.GetPlanCtx(), planDigest)
if err != nil {
return err
}
if query == "" {
return errors.NewNoStackErrorf("can't find any plans for '%s'", planDigest)
}
p := parser.New()
originNode, err := p.ParseOneStmt(query, characterSet, collation)
if err != nil {
return errors.NewNoStackErrorf("failed to parse SQL for Plan Digest: %v", planDigest)
}
hintedSQL := bindinfo.GenerateBindingSQL(originNode, planHint, schema)
stmt, err = p.ParseOneStmt(hintedSQL, characterSet, collation)
return err
})
})
return
}
func (b *PlanBuilder) buildExplain(ctx context.Context, explain *ast.ExplainStmt) (base.Plan, error) {
if show, ok := explain.Stmt.(*ast.ShowStmt); ok {
return b.buildShow(ctx, show)
}
if explain.PlanDigest != "" { // explain [analyze] <SQLDigest>
hintedStmt, err := getHintedStmtThroughPlanDigest(b.ctx, explain.PlanDigest)
if err != nil {
return nil, err
}
explain.Stmt = hintedStmt
}
sctx, err := AsSctx(b.ctx)
if err != nil {
return nil, err
}
stmtCtx := sctx.GetSessionVars().StmtCtx
var targetPlan base.Plan
if explain.Stmt != nil && !explain.Explore {
nodeW := resolve.NewNodeWWithCtx(explain.Stmt, b.resolveCtx)
if stmtCtx.ExplainFormat == types.ExplainFormatPlanCache {
targetPlan, _, err = OptimizeAstNode(ctx, sctx, nodeW, b.is)
} else {
targetPlan, _, err = OptimizeAstNodeNoCache(ctx, sctx, nodeW, b.is)
}
if err != nil {
return nil, err
}
}
return b.buildExplainPlan(targetPlan, explain.Format, "", explain.Analyze, explain.Explore, explain.Stmt, nil, explain.SQLDigest)
}
func (b *PlanBuilder) buildSelectInto(ctx context.Context, sel *ast.SelectStmt) (base.Plan, error) {
if semv1.IsEnabled() {
return nil, plannererrors.ErrNotSupportedWithSem.GenWithStackByArgs("SELECT INTO")
}
selectIntoInfo := sel.SelectIntoOpt
sel.SelectIntoOpt = nil
sctx, err := AsSctx(b.ctx)
if err != nil {
return nil, err
}
nodeW := resolve.NewNodeWWithCtx(sel, b.resolveCtx)
targetPlan, _, err := OptimizeAstNodeNoCache(ctx, sctx, nodeW, b.is)
if err != nil {
return nil, err
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.FilePriv, "", "", "", plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs("FILE"))
return &SelectInto{
TargetPlan: targetPlan,
IntoOpt: selectIntoInfo,
LineFieldsInfo: NewLineFieldsInfo(selectIntoInfo.FieldsInfo, selectIntoInfo.LinesInfo),
}, nil
}
func buildShowProcedureSchema() (*expression.Schema, []*types.FieldName) {
tblName := "ROUTINES"
schema := newColumnsWithNames(11)
schema.Append(buildColumnWithName(tblName, "Db", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Name", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Type", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Definer", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Modified", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "Created", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "Security_type", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Comment", mysql.TypeBlob, 196605))
schema.Append(buildColumnWithName(tblName, "character_set_client", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "collation_connection", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "Database Collation", mysql.TypeVarchar, 32))
return schema.col2Schema(), schema.names
}
func buildShowTriggerSchema() (*expression.Schema, []*types.FieldName) {
tblName := "TRIGGERS"
schema := newColumnsWithNames(11)
schema.Append(buildColumnWithName(tblName, "Trigger", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Event", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Table", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Statement", mysql.TypeBlob, 196605))
schema.Append(buildColumnWithName(tblName, "Timing", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Created", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "sql_mode", mysql.TypeBlob, 8192))
schema.Append(buildColumnWithName(tblName, "Definer", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "character_set_client", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "collation_connection", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "Database Collation", mysql.TypeVarchar, 32))
return schema.col2Schema(), schema.names
}
func buildShowEventsSchema() (*expression.Schema, []*types.FieldName) {
tblName := "EVENTS"
schema := newColumnsWithNames(15)
schema.Append(buildColumnWithName(tblName, "Db", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Name", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Time zone", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "Definer", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Type", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Execute At", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "Interval Value", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Interval Field", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName(tblName, "Starts", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "Ends", mysql.TypeDatetime, 19))
schema.Append(buildColumnWithName(tblName, "Status", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "Originator", mysql.TypeInt24, 4))
schema.Append(buildColumnWithName(tblName, "character_set_client", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "collation_connection", mysql.TypeVarchar, 32))
schema.Append(buildColumnWithName(tblName, "Database Collation", mysql.TypeVarchar, 32))
return schema.col2Schema(), schema.names
}
func buildShowWarningsSchema() (*expression.Schema, types.NameSlice) {
tblName := "WARNINGS"
schema := newColumnsWithNames(3)
schema.Append(buildColumnWithName(tblName, "Level", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName(tblName, "Code", mysql.TypeLong, 19))
schema.Append(buildColumnWithName(tblName, "Message", mysql.TypeVarchar, 64))
return schema.col2Schema(), schema.names
}
// buildShowSchema builds column info for ShowStmt including column name and type.
func buildShowSchema(s *ast.ShowStmt, isView bool, isSequence bool) (schema *expression.Schema, outputNames []*types.FieldName) {
var names []string
var ftypes []byte
var flags []uint
switch s.Tp {
case ast.ShowBinlogStatus:
names = []string{"File", "Position", "Binlog_Do_DB", "Binlog_Ignore_DB", "Executed_Gtid_Set"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowProcedureStatus, ast.ShowFunctionStatus:
return buildShowProcedureSchema()
case ast.ShowTriggers:
return buildShowTriggerSchema()
case ast.ShowEvents:
return buildShowEventsSchema()
case ast.ShowWarnings, ast.ShowErrors:
if s.CountWarningsOrErrors {
names = []string{"Count"}
ftypes = []byte{mysql.TypeLong}
break
}
return buildShowWarningsSchema()
case ast.ShowRegions:
return buildTableRegionsSchema()
case ast.ShowDistributions:
return buildTableDistributionSchema()
case ast.ShowEngines:
names = []string{"Engine", "Support", "Comment", "Transactions", "XA", "Savepoints"}
case ast.ShowConfig:
names = []string{"Type", "Instance", "Name", "Value"}
case ast.ShowDatabases:
fieldDB := "Database"
if patternName := extractPatternLikeOrIlikeName(s.Pattern); patternName != "" {
fieldDB = fmt.Sprintf("%s (%s)", fieldDB, patternName)
}
names = []string{fieldDB}
case ast.ShowOpenTables:
names = []string{"Database", "Table", "In_use", "Name_locked"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLong, mysql.TypeLong}
case ast.ShowTables:
fieldTable := fmt.Sprintf("Tables_in_%s", s.DBName)
if patternName := extractPatternLikeOrIlikeName(s.Pattern); patternName != "" {
fieldTable = fmt.Sprintf("%s (%s)", fieldTable, patternName)
}
names = []string{fieldTable}
if s.Full {
names = append(names, "Table_type")
}
case ast.ShowTableStatus:
names = []string{"Name", "Engine", "Version", "Row_format", "Rows", "Avg_row_length",
"Data_length", "Max_data_length", "Index_length", "Data_free", "Auto_increment",
"Create_time", "Update_time", "Check_time", "Collation", "Checksum",
"Create_options", "Comment"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeLonglong,
mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong,
mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeVarchar, mysql.TypeVarchar,
mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowColumns:
names = table.ColDescFieldNames(s.Full)
case ast.ShowCharset:
names = []string{"Charset", "Description", "Default collation", "Maxlen"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong}
case ast.ShowVariables, ast.ShowStatus:
names = []string{"Variable_name", "Value"}
case ast.ShowCollation:
names = []string{"Collation", "Charset", "Id", "Default", "Compiled", "Sortlen", "Pad_attribute"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong,
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar}
flags = []uint{0, 0, mysql.UnsignedFlag | mysql.NotNullFlag, 0, 0, 0, 0}
case ast.ShowCreateTable, ast.ShowCreateSequence:
if isSequence {
names = []string{"Sequence", "Create Sequence"}
} else if isView {
names = []string{"View", "Create View", "character_set_client", "collation_connection"}
} else {
names = []string{"Table", "Create Table"}
}
case ast.ShowCreatePlacementPolicy:
names = []string{"Policy", "Create Policy"}
case ast.ShowCreateResourceGroup:
names = []string{"Resource_Group", "Create Resource Group"}
case ast.ShowCreateUser:
if s.User != nil {
names = []string{fmt.Sprintf("CREATE USER for %s", s.User)}
}
case ast.ShowCreateView:
names = []string{"View", "Create View", "character_set_client", "collation_connection"}
case ast.ShowCreateDatabase:
names = []string{"Database", "Create Database"}
case ast.ShowGrants:
if s.User != nil {
names = []string{fmt.Sprintf("Grants for %s", s.User)}
} else {
// Don't know the name yet, so just say "user"
names = []string{"Grants for User"}
}
case ast.ShowIndex:
names = []string{"Table", "Non_unique", "Key_name", "Seq_in_index",
"Column_name", "Collation", "Cardinality", "Sub_part", "Packed",
"Null", "Index_type", "Comment", "Index_comment", "Visible", "Expression", "Clustered", "Global"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeLonglong,
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeLonglong,
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar,
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowPlugins:
names = []string{"Name", "Status", "Type", "Library", "License", "Version"}
ftypes = []byte{
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar,
}
case ast.ShowProcessList:
names = []string{"Id", "User", "Host", "db", "Command", "Time", "State", "Info"}
ftypes = []byte{mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar,
mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLong, mysql.TypeVarchar, mysql.TypeString}
case ast.ShowStatsMeta:
names = []string{"Db_name", "Table_name", "Partition_name", "Update_time", "Modify_count", "Row_count", "Last_analyze_time"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeDatetime, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeDatetime}
case ast.ShowStatsExtended:
names = []string{"Db_name", "Table_name", "Stats_name", "Column_names", "Stats_type", "Stats_val", "Last_update_version"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong}
case ast.ShowStatsHistograms:
names = []string{"Db_name", "Table_name", "Partition_name", "Column_name", "Is_index", "Update_time", "Distinct_count", "Null_count", "Avg_col_size", "Correlation",
"Load_status", "Total_mem_usage", "Hist_mem_usage", "Topn_mem_usage", "Cms_mem_usage"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeTiny, mysql.TypeDatetime,
mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeDouble, mysql.TypeDouble,
mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeLonglong}
case ast.ShowStatsBuckets:
names = []string{"Db_name", "Table_name", "Partition_name", "Column_name", "Is_index", "Bucket_id", "Count",
"Repeats", "Lower_Bound", "Upper_Bound", "Ndv"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeTiny, mysql.TypeLonglong,
mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong}
case ast.ShowStatsTopN:
names = []string{"Db_name", "Table_name", "Partition_name", "Column_name", "Is_index", "Value", "Count"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeTiny, mysql.TypeVarchar, mysql.TypeLonglong}
case ast.ShowStatsHealthy:
names = []string{"Db_name", "Table_name", "Partition_name", "Healthy"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong}
case ast.ShowHistogramsInFlight:
names = []string{"HistogramsInFlight"}
ftypes = []byte{mysql.TypeLonglong}
case ast.ShowColumnStatsUsage:
names = []string{"Db_name", "Table_name", "Partition_name", "Column_name", "Last_used_at", "Last_analyzed_at"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeDatetime, mysql.TypeDatetime}
case ast.ShowStatsLocked:
names = []string{"Db_name", "Table_name", "Partition_name", "Status"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowProfiles: // ShowProfiles is deprecated.
names = []string{"Query_ID", "Duration", "Query"}
ftypes = []byte{mysql.TypeLong, mysql.TypeDouble, mysql.TypeVarchar}
case ast.ShowMasterStatus:
names = []string{"File", "Position", "Binlog_Do_DB", "Binlog_Ignore_DB", "Executed_Gtid_Set"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowPrivileges:
names = []string{"Privilege", "Context", "Comment"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowBindings:
names = []string{"Original_sql", "Bind_sql", "Default_db", "Status", "Create_time", "Update_time", "Charset", "Collation", "Source", "Sql_digest", "Plan_digest"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowBindingCacheStatus:
names = []string{"bindings_in_cache", "bindings_in_table", "memory_usage", "memory_quota"}
ftypes = []byte{mysql.TypeLonglong, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowAnalyzeStatus:
names = []string{"Table_schema", "Table_name", "Partition_name", "Job_info", "Processed_rows", "Start_time",
"End_time", "State", "Fail_reason", "Instance", "Process_ID", "Remaining_seconds", "Progress", "Estimated_total_rows"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong,
mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeDouble, mysql.TypeLonglong}
case ast.ShowBuiltins:
names = []string{"Supported_builtin_functions"}
ftypes = []byte{mysql.TypeVarchar}
case ast.ShowBackups, ast.ShowRestores:
names = []string{"Id", "Destination", "State", "Progress", "Queue_time", "Execution_time", "Finish_time", "Connection", "Message"}
ftypes = []byte{mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeDouble, mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeDatetime, mysql.TypeLonglong, mysql.TypeVarchar}
case ast.ShowPlacementLabels:
names = []string{"Key", "Values"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeJSON}
case ast.ShowPlacement, ast.ShowPlacementForDatabase, ast.ShowPlacementForTable, ast.ShowPlacementForPartition:
names = []string{"Target", "Placement", "Scheduling_State"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar}
case ast.ShowSessionStates:
names = []string{"Session_states", "Session_token"}
ftypes = []byte{mysql.TypeJSON, mysql.TypeJSON}
case ast.ShowImportJobs:
names = importIntoSchemaNames
ftypes = ImportIntoSchemaFTypes
case ast.ShowImportGroups:
names = showImportGroupsNames
ftypes = showImportGroupsFTypes
case ast.ShowDistributionJobs:
names = distributionJobsSchemaNames
ftypes = distributionJobsSchedulerFTypes
case ast.ShowAffinity:
names = []string{"Db_name", "Table_name", "Partition_name", "Leader_store_id", "Voter_store_ids", "Status", "Region_count", "Affinity_region_count"}
ftypes = []byte{mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeVarchar, mysql.TypeVarchar, mysql.TypeLonglong, mysql.TypeLonglong}
}
return convert2OutputSchemasAndNames(names, ftypes, flags)
}
func convert2OutputSchemasAndNames(names []string, ftypes []byte, flags []uint) (schema *expression.Schema, outputNames []*types.FieldName) {
schema = expression.NewSchema(make([]*expression.Column, 0, len(names))...)
outputNames = make([]*types.FieldName, 0, len(names))
for i := range names {
col := &expression.Column{}
outputNames = append(outputNames, &types.FieldName{ColName: ast.NewCIStr(names[i])})
// User varchar as the default return column type.
tp := mysql.TypeVarchar
if len(ftypes) != 0 && ftypes[i] != mysql.TypeUnspecified {
tp = ftypes[i]
}
fieldType := types.NewFieldType(tp)
flen, decimal := mysql.GetDefaultFieldLengthAndDecimal(tp)
fieldType.SetFlen(flen)
fieldType.SetDecimal(decimal)
charset, collate := types.DefaultCharsetForType(tp)
fieldType.SetCharset(charset)
fieldType.SetCollate(collate)
if len(flags) > 0 {
fieldType.SetFlag(flags[i])
}
col.RetType = fieldType
schema.Append(col)
}
return
}
func (b *PlanBuilder) buildPlanReplayer(pc *ast.PlanReplayerStmt) base.Plan {
p := &PlanReplayer{ExecStmt: pc.Stmt, Analyze: pc.Analyze, Load: pc.Load, File: pc.File,
Capture: pc.Capture, Remove: pc.Remove, SQLDigest: pc.SQLDigest, PlanDigest: pc.PlanDigest}
if pc.HistoricalStatsInfo != nil {
p.HistoricalStatsTS = calcTSForPlanReplayer(b.ctx, pc.HistoricalStatsInfo.TsExpr)
}
schema := newColumnsWithNames(1)
schema.Append(buildColumnWithName("", "File_token", mysql.TypeVarchar, 128))
p.SetSchema(schema.col2Schema())
p.SetOutputNames(schema.names)
return p
}
func calcTSForPlanReplayer(sctx base.PlanContext, tsExpr ast.ExprNode) uint64 {
tsVal, err := evalAstExprWithPlanCtx(sctx, tsExpr)
if err != nil {
sctx.GetSessionVars().StmtCtx.AppendWarning(err)
return 0
}
// mustn't be NULL
if tsVal.IsNull() {
return 0
}
// first, treat it as a TSO
tpLonglong := types.NewFieldType(mysql.TypeLonglong)
tpLonglong.SetFlag(mysql.UnsignedFlag)
// We need a strict check, which means no truncate or any other warnings/errors, or it will wrongly try to parse
// a date/time string into a TSO.
// To achieve this, we create a new type context without re-using the one in statement context.
tso, err := tsVal.ConvertTo(types.DefaultStmtNoWarningContext.WithLocation(sctx.GetSessionVars().Location()), tpLonglong)
if err == nil {
return tso.GetUint64()
}
// if failed, treat it as a date/time
// this part is similar to CalculateAsOfTsExpr
tpDateTime := types.NewFieldType(mysql.TypeDatetime)
tpDateTime.SetDecimal(6)
timestamp, err := tsVal.ConvertTo(sctx.GetSessionVars().StmtCtx.TypeCtx(), tpDateTime)
if err != nil {
sctx.GetSessionVars().StmtCtx.AppendWarning(err)
return 0
}
goTime, err := timestamp.GetMysqlTime().GoTime(sctx.GetSessionVars().Location())
if err != nil {
sctx.GetSessionVars().StmtCtx.AppendWarning(err)
return 0
}
return oracle.GoTimeToTS(goTime)
}
func buildChecksumTableSchema() (*expression.Schema, []*types.FieldName) {
schema := newColumnsWithNames(5)
schema.Append(buildColumnWithName("", "Db_name", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName("", "Table_name", mysql.TypeVarchar, 128))
schema.Append(buildColumnWithName("", "Checksum_crc64_xor", mysql.TypeLonglong, 22))
schema.Append(buildColumnWithName("", "Total_kvs", mysql.TypeLonglong, 22))
schema.Append(buildColumnWithName("", "Total_bytes", mysql.TypeLonglong, 22))
return schema.col2Schema(), schema.names
}
// adjustOverlongViewColname adjusts the overlong outputNames of a view to
// `new_exp_$off` where `$off` is the offset of the output column, $off starts from 1.
// There is still some MySQL compatible problems.
func adjustOverlongViewColname(plan base.LogicalPlan) {
outputNames := plan.OutputNames()
for i := range outputNames {
if outputName := outputNames[i].ColName.L; len(outputName) > mysql.MaxColumnNameLength {
outputNames[i].ColName = ast.NewCIStr(fmt.Sprintf("name_exp_%d", i+1))
}
}
}
// findStmtAsViewSchema finds the first SelectStmt as the schema for the view
func findStmtAsViewSchema(stmt ast.Node) *ast.SelectStmt {
switch x := stmt.(type) {
case *ast.CreateViewStmt:
return findStmtAsViewSchema(x.Select)
case *ast.SetOprStmt:
return findStmtAsViewSchema(x.SelectList)
case *ast.SetOprSelectList:
return findStmtAsViewSchema(x.Selects[0])
case *ast.SelectStmt:
return x
}
return nil
}
func (b *PlanBuilder) buildTraffic(pc *ast.TrafficStmt) base.Plan {
tf := &Traffic{
OpType: pc.OpType,
Options: pc.Options,
Dir: pc.Dir,
}
var errMsg string
var privs []string
switch pc.OpType {
case ast.TrafficOpCapture:
errMsg = "SUPER or TRAFFIC_CAPTURE_ADMIN"
privs = []string{"TRAFFIC_CAPTURE_ADMIN"}
case ast.TrafficOpReplay:
errMsg = "SUPER or TRAFFIC_REPLAY_ADMIN"
privs = []string{"TRAFFIC_REPLAY_ADMIN"}
case ast.TrafficOpShow:
tf.SetSchemaAndNames(buildShowTrafficJobsSchema())
fallthrough
case ast.TrafficOpCancel:
errMsg = "SUPER, TRAFFIC_CAPTURE_ADMIN or TRAFFIC_REPLAY_ADMIN"
privs = []string{"TRAFFIC_CAPTURE_ADMIN", "TRAFFIC_REPLAY_ADMIN"}
}
err := plannererrors.ErrSpecificAccessDenied.GenWithStackByArgs(errMsg)
b.visitInfo = appendDynamicVisitInfo(b.visitInfo, privs, false, err)
return tf
}
// buildCompactTable builds a plan for the "ALTER TABLE [NAME] COMPACT ..." statement.
func (b *PlanBuilder) buildCompactTable(node *ast.CompactTableStmt) (base.Plan, error) {
var authErr error
if b.ctx.GetSessionVars().User != nil {
authErr = plannererrors.ErrTableaccessDenied.GenWithStackByArgs("ALTER", b.ctx.GetSessionVars().User.AuthUsername,
b.ctx.GetSessionVars().User.AuthHostname, node.Table.Name.L)
}
b.visitInfo = appendVisitInfo(b.visitInfo, mysql.AlterPriv, node.Table.Schema.L,
node.Table.Name.L, "", authErr)
tnW := b.resolveCtx.GetTableName(node.Table)
tblInfo := tnW.TableInfo
p := &CompactTable{
ReplicaKind: node.ReplicaKind,
TableInfo: tblInfo,
PartitionNames: node.PartitionNames,
}
return p, nil
}
func (*PlanBuilder) buildRecommendIndex(v *ast.RecommendIndexStmt) (base.Plan, error) {
p := &RecommendIndexPlan{
Action: v.Action,
SQL: v.SQL,
AdviseID: v.ID,
Options: v.Options,
}
switch v.Action {
case "run":
schema := newColumnsWithNames(7)
schema.Append(buildColumnWithName("", "database", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "table", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "index_name", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "index_columns", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "est_index_size", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "reason", mysql.TypeVarchar, 256))
schema.Append(buildColumnWithName("", "top_impacted_query", mysql.TypeBlob, -1))
schema.Append(buildColumnWithName("", "create_index_statement", mysql.TypeBlob, -1))
p.SetSchemaAndNames(schema.col2Schema(), schema.names)
case "set":
if len(p.Options) == 0 {
return nil, fmt.Errorf("option is empty")
}
case "show":
schema := newColumnsWithNames(3)
schema.Append(buildColumnWithName("", "option", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "value", mysql.TypeVarchar, 64))
schema.Append(buildColumnWithName("", "description", mysql.TypeVarchar, 256))
p.SetSchemaAndNames(schema.col2Schema(), schema.names)
default:
return nil, fmt.Errorf("unsupported action %s", v.Action)
}
return p, nil
}
func extractPatternLikeOrIlikeName(patternLike *ast.PatternLikeOrIlikeExpr) string {
if patternLike == nil {
return ""
}
if v, ok := patternLike.Pattern.(*driver.ValueExpr); ok {
return v.GetString()
}
return ""
}
// getTablePath finds the TablePath from a group of accessPaths.
func getTablePath(paths []*util.AccessPath) *util.AccessPath {
for _, path := range paths {
if path.IsTablePath() {
return path
}
}
return nil
}
func (b *PlanBuilder) buildAdminAlterDDLJob(ctx context.Context, as *ast.AdminStmt) (_ base.Plan, err error) {
options := make([]*AlterDDLJobOpt, 0, len(as.AlterJobOptions))
mockTablePlan := logicalop.LogicalTableDual{}.Init(b.ctx, b.getSelectOffset())
for _, opt := range as.AlterJobOptions {
_, ok := allowedAlterDDLJobParams[opt.Name]
if !ok {
return nil, fmt.Errorf("unsupported admin alter ddl jobs config: %s", opt.Name)
}
alterDDLJobOpt := AlterDDLJobOpt{Name: opt.Name}
if opt.Value != nil {
alterDDLJobOpt.Value, _, err = b.rewrite(ctx, opt.Value, mockTablePlan, nil, true)
if err != nil {
return nil, err
}
}
if err = checkAlterDDLJobOptValue(&alterDDLJobOpt); err != nil {
return nil, err
}
options = append(options, &alterDDLJobOpt)
}
p := &AlterDDLJob{
JobID: as.JobNumber,
Options: options,
}
return p, nil
}
// check if the config value is valid.
func checkAlterDDLJobOptValue(opt *AlterDDLJobOpt) error {
switch opt.Name {
case AlterDDLJobThread:
thread, err := GetThreadOrBatchSizeFromExpression(opt)
if err != nil {
return err
}
if thread < 1 || thread > vardef.MaxConfigurableConcurrency {
return fmt.Errorf("the value %v for %s is out of range [1, %v]",
thread, opt.Name, vardef.MaxConfigurableConcurrency)
}
case AlterDDLJobBatchSize:
batchSize, err := GetThreadOrBatchSizeFromExpression(opt)
if err != nil {
return err
}
bs := int32(batchSize)
if bs < vardef.MinDDLReorgBatchSize || bs > vardef.MaxDDLReorgBatchSize {
return fmt.Errorf("the value %v for %s is out of range [%v, %v]",
bs, opt.Name, vardef.MinDDLReorgBatchSize, vardef.MaxDDLReorgBatchSize)
}
case AlterDDLJobMaxWriteSpeed:
speed, err := GetMaxWriteSpeedFromExpression(opt)
if err != nil {
return err
}
if speed < 0 || speed > units.PiB {
return fmt.Errorf("the value %s for %s is out of range [%v, %v]",
strconv.FormatInt(speed, 10), opt.Name, 0, units.PiB)
}
}
return nil
}
// for nextgen import-into with SEM, we disallow user to set S3 external ID explicitly,
// and we will use the keyspace name as the S3 external ID.
func checkNextGenS3PathWithSem(u *url.URL) error {
values := u.Query()
for k := range values {
lowerK := strings.ToLower(k)
if lowerK == s3like.S3ExternalID {
return plannererrors.ErrNotSupportedWithSem.GenWithStackByArgs("IMPORT INTO with S3 external ID")
}
}
return nil
}
// GetThreadOrBatchSizeFromExpression gets the numeric value of the thread or batch size from the expression.
func GetThreadOrBatchSizeFromExpression(opt *AlterDDLJobOpt) (int64, error) {
v := opt.Value.(*expression.Constant)
switch v.RetType.EvalType() {
case types.ETInt:
return v.Value.GetInt64(), nil
default:
return 0, fmt.Errorf("the value for %s is invalid, only integer is allowed", opt.Name)
}
}
// GetMaxWriteSpeedFromExpression gets the numeric value of the max write speed from the expression.
func GetMaxWriteSpeedFromExpression(opt *AlterDDLJobOpt) (maxWriteSpeed int64, err error) {
v := opt.Value.(*expression.Constant)
switch v.RetType.EvalType() {
case types.ETString:
speedStr := v.Value.GetString()
maxWriteSpeed, err = units.RAMInBytes(speedStr)
if err != nil {
return 0, errors.Annotate(err, "parse max_write_speed value error")
}
case types.ETInt:
maxWriteSpeed = v.Value.GetInt64()
default:
return 0, fmt.Errorf("the value %v for %s is invalid", v.Value.GetValue(), opt.Name)
}
return maxWriteSpeed, nil
}
func (b *PlanBuilder) checkSEMStmt(stmt ast.Node) error {
if !semv2.IsEnabled() {
return nil
}
stmtNode, ok := stmt.(ast.StmtNode)
intest.Assert(ok, "node.Node should be ast.StmtNode, but got %T", stmt)
if !semv2.IsRestrictedSQL(stmtNode) {
return nil
}
activeRoles := b.ctx.GetSessionVars().ActiveRoles
checker := privilege.GetPrivilegeManager(b.ctx)
hasPriv := checker.RequestDynamicVerification(activeRoles, "RESTRICTED_SQL_ADMIN", false)
if hasPriv {
return nil
}
return plannererrors.ErrNotSupportedWithSem.GenWithStackByArgs(stmtNode.Text())
}