825 lines
23 KiB
Go
825 lines
23 KiB
Go
// Copyright 2018 PingCAP, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package executor
|
|
|
|
import (
|
|
"context"
|
|
"math"
|
|
|
|
"github.com/pingcap/errors"
|
|
"github.com/pingcap/kvproto/pkg/kvrpcpb"
|
|
"github.com/pingcap/tidb/distsql"
|
|
"github.com/pingcap/tidb/infoschema"
|
|
"github.com/pingcap/tidb/kv"
|
|
"github.com/pingcap/tidb/parser/ast"
|
|
"github.com/pingcap/tidb/parser/model"
|
|
"github.com/pingcap/tidb/parser/terror"
|
|
plannercore "github.com/pingcap/tidb/planner/core"
|
|
"github.com/pingcap/tidb/sessionctx/stmtctx"
|
|
"github.com/pingcap/tidb/statistics"
|
|
"github.com/pingcap/tidb/table"
|
|
"github.com/pingcap/tidb/table/tables"
|
|
"github.com/pingcap/tidb/tablecodec"
|
|
"github.com/pingcap/tidb/types"
|
|
"github.com/pingcap/tidb/util"
|
|
"github.com/pingcap/tidb/util/chunk"
|
|
"github.com/pingcap/tidb/util/logutil"
|
|
"github.com/pingcap/tidb/util/ranger"
|
|
"github.com/pingcap/tidb/util/timeutil"
|
|
"github.com/pingcap/tipb/go-tipb"
|
|
"go.uber.org/zap"
|
|
)
|
|
|
|
var (
|
|
_ Executor = &CheckIndexRangeExec{}
|
|
_ Executor = &RecoverIndexExec{}
|
|
_ Executor = &CleanupIndexExec{}
|
|
)
|
|
|
|
// CheckIndexRangeExec outputs the index values which has handle between begin and end.
|
|
type CheckIndexRangeExec struct {
|
|
baseExecutor
|
|
|
|
table *model.TableInfo
|
|
index *model.IndexInfo
|
|
is infoschema.InfoSchema
|
|
startKey []types.Datum
|
|
|
|
handleRanges []ast.HandleRange
|
|
srcChunk *chunk.Chunk
|
|
|
|
result distsql.SelectResult
|
|
cols []*model.ColumnInfo
|
|
}
|
|
|
|
// Next implements the Executor Next interface.
|
|
func (e *CheckIndexRangeExec) Next(ctx context.Context, req *chunk.Chunk) error {
|
|
req.Reset()
|
|
handleIdx := e.schema.Len() - 1
|
|
for {
|
|
err := e.result.Next(ctx, e.srcChunk)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if e.srcChunk.NumRows() == 0 {
|
|
return nil
|
|
}
|
|
iter := chunk.NewIterator4Chunk(e.srcChunk)
|
|
appendRows := make([]chunk.Row, 0, e.srcChunk.NumRows())
|
|
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
|
|
handle := row.GetInt64(handleIdx)
|
|
for _, hr := range e.handleRanges {
|
|
if handle >= hr.Begin && handle < hr.End {
|
|
appendRows = append(appendRows, row)
|
|
break
|
|
}
|
|
}
|
|
}
|
|
if len(appendRows) > 0 {
|
|
req.AppendRows(appendRows)
|
|
}
|
|
if req.NumRows() > 0 {
|
|
return nil
|
|
}
|
|
}
|
|
}
|
|
|
|
// Open implements the Executor Open interface.
|
|
func (e *CheckIndexRangeExec) Open(ctx context.Context) error {
|
|
tCols := e.table.Cols()
|
|
for _, ic := range e.index.Columns {
|
|
col := tCols[ic.Offset]
|
|
e.cols = append(e.cols, col)
|
|
}
|
|
|
|
colTypeForHandle := e.schema.Columns[len(e.cols)].RetType
|
|
e.cols = append(e.cols, &model.ColumnInfo{
|
|
ID: model.ExtraHandleID,
|
|
Name: model.ExtraHandleName,
|
|
FieldType: *colTypeForHandle,
|
|
})
|
|
|
|
e.srcChunk = tryNewCacheChunk(e)
|
|
dagPB, err := e.buildDAGPB()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
txn, err := e.ctx.Txn(true)
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
var builder distsql.RequestBuilder
|
|
kvReq, err := builder.SetIndexRanges(sc, e.table.ID, e.index.ID, ranger.FullRange()).
|
|
SetDAGRequest(dagPB).
|
|
SetStartTS(txn.StartTS()).
|
|
SetKeepOrder(true).
|
|
SetFromSessionVars(e.ctx.GetSessionVars()).
|
|
SetFromInfoSchema(e.ctx.GetInfoSchema()).
|
|
Build()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
e.result, err = distsql.Select(ctx, e.ctx, kvReq, e.retFieldTypes, statistics.NewQueryFeedback(0, nil, 0, false))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (e *CheckIndexRangeExec) buildDAGPB() (*tipb.DAGRequest, error) {
|
|
dagReq := &tipb.DAGRequest{}
|
|
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
dagReq.Flags = sc.PushDownFlags()
|
|
for i := range e.schema.Columns {
|
|
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
|
|
}
|
|
execPB := e.constructIndexScanPB()
|
|
dagReq.Executors = append(dagReq.Executors, execPB)
|
|
|
|
err := tables.SetPBColumnsDefaultValue(e.ctx, dagReq.Executors[0].IdxScan.Columns, e.cols)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
distsql.SetEncodeType(e.ctx, dagReq)
|
|
return dagReq, nil
|
|
}
|
|
|
|
func (e *CheckIndexRangeExec) constructIndexScanPB() *tipb.Executor {
|
|
idxExec := &tipb.IndexScan{
|
|
TableId: e.table.ID,
|
|
IndexId: e.index.ID,
|
|
Columns: util.ColumnsToProto(e.cols, e.table.PKIsHandle),
|
|
}
|
|
return &tipb.Executor{Tp: tipb.ExecType_TypeIndexScan, IdxScan: idxExec}
|
|
}
|
|
|
|
// Close implements the Executor Close interface.
|
|
func (e *CheckIndexRangeExec) Close() error {
|
|
return nil
|
|
}
|
|
|
|
// RecoverIndexExec represents a recover index executor.
|
|
// It is built from "admin recover index" statement, is used to backfill
|
|
// corrupted index.
|
|
type RecoverIndexExec struct {
|
|
baseExecutor
|
|
|
|
done bool
|
|
|
|
index table.Index
|
|
table table.Table
|
|
physicalID int64
|
|
batchSize int
|
|
|
|
columns []*model.ColumnInfo
|
|
colFieldTypes []*types.FieldType
|
|
srcChunk *chunk.Chunk
|
|
handleCols plannercore.HandleCols
|
|
|
|
// below buf is used to reduce allocations.
|
|
recoverRows []recoverRows
|
|
idxValsBufs [][]types.Datum
|
|
idxKeyBufs [][]byte
|
|
batchKeys []kv.Key
|
|
}
|
|
|
|
func (e *RecoverIndexExec) columnsTypes() []*types.FieldType {
|
|
if e.colFieldTypes != nil {
|
|
return e.colFieldTypes
|
|
}
|
|
|
|
e.colFieldTypes = make([]*types.FieldType, 0, len(e.columns))
|
|
for _, col := range e.columns {
|
|
e.colFieldTypes = append(e.colFieldTypes, &col.FieldType)
|
|
}
|
|
return e.colFieldTypes
|
|
}
|
|
|
|
// Open implements the Executor Open interface.
|
|
func (e *RecoverIndexExec) Open(ctx context.Context) error {
|
|
if err := e.baseExecutor.Open(ctx); err != nil {
|
|
return err
|
|
}
|
|
|
|
e.srcChunk = chunk.New(e.columnsTypes(), e.initCap, e.maxChunkSize)
|
|
e.batchSize = 2048
|
|
e.recoverRows = make([]recoverRows, 0, e.batchSize)
|
|
e.idxValsBufs = make([][]types.Datum, e.batchSize)
|
|
e.idxKeyBufs = make([][]byte, e.batchSize)
|
|
return nil
|
|
}
|
|
|
|
func (e *RecoverIndexExec) constructTableScanPB(tblInfo *model.TableInfo, colInfos []*model.ColumnInfo) (*tipb.Executor, error) {
|
|
tblScan := tables.BuildTableScanFromInfos(tblInfo, colInfos)
|
|
tblScan.TableId = e.physicalID
|
|
err := tables.SetPBColumnsDefaultValue(e.ctx, tblScan.Columns, colInfos)
|
|
return &tipb.Executor{Tp: tipb.ExecType_TypeTableScan, TblScan: tblScan}, err
|
|
}
|
|
|
|
func (e *RecoverIndexExec) constructLimitPB(count uint64) *tipb.Executor {
|
|
limitExec := &tipb.Limit{
|
|
Limit: count,
|
|
}
|
|
return &tipb.Executor{Tp: tipb.ExecType_TypeLimit, Limit: limitExec}
|
|
}
|
|
|
|
func (e *RecoverIndexExec) buildDAGPB(txn kv.Transaction, limitCnt uint64) (*tipb.DAGRequest, error) {
|
|
dagReq := &tipb.DAGRequest{}
|
|
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
dagReq.Flags = sc.PushDownFlags()
|
|
for i := range e.columns {
|
|
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
|
|
}
|
|
|
|
tblScanExec, err := e.constructTableScanPB(e.table.Meta(), e.columns)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
dagReq.Executors = append(dagReq.Executors, tblScanExec)
|
|
|
|
limitExec := e.constructLimitPB(limitCnt)
|
|
dagReq.Executors = append(dagReq.Executors, limitExec)
|
|
distsql.SetEncodeType(e.ctx, dagReq)
|
|
return dagReq, nil
|
|
}
|
|
|
|
func (e *RecoverIndexExec) buildTableScan(ctx context.Context, txn kv.Transaction, startHandle kv.Handle, limitCnt uint64) (distsql.SelectResult, error) {
|
|
dagPB, err := e.buildDAGPB(txn, limitCnt)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
var builder distsql.RequestBuilder
|
|
builder.KeyRanges, err = buildRecoverIndexKeyRanges(e.ctx.GetSessionVars().StmtCtx, e.physicalID, startHandle)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
kvReq, err := builder.
|
|
SetDAGRequest(dagPB).
|
|
SetStartTS(txn.StartTS()).
|
|
SetKeepOrder(true).
|
|
SetFromSessionVars(e.ctx.GetSessionVars()).
|
|
SetFromInfoSchema(e.ctx.GetInfoSchema()).
|
|
Build()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Actually, with limitCnt, the match datas maybe only in one region, so let the concurrency to be 1,
|
|
// avoid unnecessary region scan.
|
|
kvReq.Concurrency = 1
|
|
result, err := distsql.Select(ctx, e.ctx, kvReq, e.columnsTypes(), statistics.NewQueryFeedback(0, nil, 0, false))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// buildRecoverIndexKeyRanges build a KeyRange: (startHandle, unlimited).
|
|
func buildRecoverIndexKeyRanges(sctx *stmtctx.StatementContext, tid int64, startHandle kv.Handle) ([]kv.KeyRange, error) {
|
|
var startKey []byte
|
|
if startHandle == nil {
|
|
startKey = tablecodec.GenTableRecordPrefix(tid).Next()
|
|
} else {
|
|
startKey = tablecodec.EncodeRowKey(tid, startHandle.Encoded()).PrefixNext()
|
|
}
|
|
endKey := tablecodec.GenTableRecordPrefix(tid).PrefixNext()
|
|
return []kv.KeyRange{{StartKey: startKey, EndKey: endKey}}, nil
|
|
}
|
|
|
|
type backfillResult struct {
|
|
currentHandle kv.Handle
|
|
addedCount int64
|
|
scanRowCount int64
|
|
}
|
|
|
|
func (e *RecoverIndexExec) backfillIndex(ctx context.Context) (int64, int64, error) {
|
|
var (
|
|
currentHandle kv.Handle = nil
|
|
totalAddedCnt = int64(0)
|
|
totalScanCnt = int64(0)
|
|
lastLogCnt = int64(0)
|
|
result backfillResult
|
|
)
|
|
for {
|
|
ctx = kv.WithInternalSourceType(ctx, kv.InternalTxnAdmin)
|
|
errInTxn := kv.RunInNewTxn(ctx, e.ctx.GetStore(), true, func(ctx context.Context, txn kv.Transaction) error {
|
|
setOptionForTopSQL(e.ctx.GetSessionVars().StmtCtx, txn)
|
|
var err error
|
|
result, err = e.backfillIndexInTxn(ctx, txn, currentHandle)
|
|
return err
|
|
})
|
|
if errInTxn != nil {
|
|
return totalAddedCnt, totalScanCnt, errInTxn
|
|
}
|
|
totalAddedCnt += result.addedCount
|
|
totalScanCnt += result.scanRowCount
|
|
if totalScanCnt-lastLogCnt >= 50000 {
|
|
lastLogCnt = totalScanCnt
|
|
logutil.Logger(ctx).Info("recover index", zap.String("table", e.table.Meta().Name.O),
|
|
zap.String("index", e.index.Meta().Name.O), zap.Int64("totalAddedCnt", totalAddedCnt),
|
|
zap.Int64("totalScanCnt", totalScanCnt), zap.Stringer("currentHandle", result.currentHandle))
|
|
}
|
|
|
|
// no more rows
|
|
if result.scanRowCount == 0 {
|
|
break
|
|
}
|
|
currentHandle = result.currentHandle
|
|
if currentHandle.Next().Compare(result.currentHandle) <= 0 {
|
|
break // There is no more handles in the table.
|
|
}
|
|
}
|
|
return totalAddedCnt, totalScanCnt, nil
|
|
}
|
|
|
|
type recoverRows struct {
|
|
handle kv.Handle
|
|
idxVals []types.Datum
|
|
rsData []types.Datum
|
|
skip bool
|
|
}
|
|
|
|
func (e *RecoverIndexExec) fetchRecoverRows(ctx context.Context, srcResult distsql.SelectResult, result *backfillResult) ([]recoverRows, error) {
|
|
e.recoverRows = e.recoverRows[:0]
|
|
idxValLen := len(e.index.Meta().Columns)
|
|
result.scanRowCount = 0
|
|
|
|
for {
|
|
err := srcResult.Next(ctx, e.srcChunk)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if e.srcChunk.NumRows() == 0 {
|
|
break
|
|
}
|
|
iter := chunk.NewIterator4Chunk(e.srcChunk)
|
|
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
|
|
if result.scanRowCount >= int64(e.batchSize) {
|
|
return e.recoverRows, nil
|
|
}
|
|
handle, err := e.handleCols.BuildHandle(row)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
idxVals := extractIdxVals(row, e.idxValsBufs[result.scanRowCount], e.colFieldTypes, idxValLen)
|
|
e.idxValsBufs[result.scanRowCount] = idxVals
|
|
rsData := tables.TryGetHandleRestoredDataWrapper(e.table.Meta(), plannercore.GetCommonHandleDatum(e.handleCols, row), nil, e.index.Meta())
|
|
e.recoverRows = append(e.recoverRows, recoverRows{handle: handle, idxVals: idxVals, rsData: rsData, skip: false})
|
|
result.scanRowCount++
|
|
result.currentHandle = handle
|
|
}
|
|
}
|
|
|
|
return e.recoverRows, nil
|
|
}
|
|
|
|
func (e *RecoverIndexExec) batchMarkDup(txn kv.Transaction, rows []recoverRows) error {
|
|
if len(rows) == 0 {
|
|
return nil
|
|
}
|
|
e.batchKeys = e.batchKeys[:0]
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
distinctFlags := make([]bool, len(rows))
|
|
for i, row := range rows {
|
|
idxKey, distinct, err := e.index.GenIndexKey(sc, row.idxVals, row.handle, e.idxKeyBufs[i])
|
|
if err != nil {
|
|
return err
|
|
}
|
|
e.idxKeyBufs[i] = idxKey
|
|
|
|
e.batchKeys = append(e.batchKeys, idxKey)
|
|
distinctFlags[i] = distinct
|
|
}
|
|
|
|
values, err := txn.BatchGet(context.Background(), e.batchKeys)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// 1. unique-key is duplicate and the handle is equal, skip it.
|
|
// 2. unique-key is duplicate and the handle is not equal, data is not consistent, log it and skip it.
|
|
// 3. non-unique-key is duplicate, skip it.
|
|
isCommonHandle := e.table.Meta().IsCommonHandle
|
|
for i, key := range e.batchKeys {
|
|
if val, found := values[string(key)]; found {
|
|
if distinctFlags[i] {
|
|
handle, err1 := tablecodec.DecodeHandleInUniqueIndexValue(val, isCommonHandle)
|
|
if err1 != nil {
|
|
return err1
|
|
}
|
|
|
|
if handle.Compare(rows[i].handle) != 0 {
|
|
logutil.BgLogger().Warn("recover index: the constraint of unique index is broken, handle in index is not equal to handle in table",
|
|
zap.String("index", e.index.Meta().Name.O), zap.ByteString("indexKey", key),
|
|
zap.Stringer("handleInTable", rows[i].handle), zap.Stringer("handleInIndex", handle))
|
|
}
|
|
}
|
|
rows[i].skip = true
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (e *RecoverIndexExec) backfillIndexInTxn(ctx context.Context, txn kv.Transaction, currentHandle kv.Handle) (result backfillResult, err error) {
|
|
srcResult, err := e.buildTableScan(ctx, txn, currentHandle, uint64(e.batchSize))
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
defer terror.Call(srcResult.Close)
|
|
|
|
rows, err := e.fetchRecoverRows(ctx, srcResult, &result)
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
|
|
err = e.batchMarkDup(txn, rows)
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
|
|
// Constrains is already checked.
|
|
e.ctx.GetSessionVars().StmtCtx.BatchCheck = true
|
|
for _, row := range rows {
|
|
if row.skip {
|
|
continue
|
|
}
|
|
|
|
recordKey := tablecodec.EncodeRecordKey(e.table.RecordPrefix(), row.handle)
|
|
err := txn.LockKeys(ctx, new(kv.LockCtx), recordKey)
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
|
|
_, err = e.index.Create(e.ctx, txn, row.idxVals, row.handle, row.rsData, table.WithIgnoreAssertion)
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
result.addedCount++
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Next implements the Executor Next interface.
|
|
func (e *RecoverIndexExec) Next(ctx context.Context, req *chunk.Chunk) error {
|
|
req.Reset()
|
|
if e.done {
|
|
return nil
|
|
}
|
|
|
|
recoveringClusteredIndex := e.index.Meta().Primary && e.table.Meta().IsCommonHandle
|
|
if recoveringClusteredIndex {
|
|
req.AppendInt64(0, 0)
|
|
req.AppendInt64(1, 0)
|
|
e.done = true
|
|
return nil
|
|
}
|
|
var totalAddedCnt, totalScanCnt int64
|
|
var err error
|
|
if tbl, ok := e.table.(table.PartitionedTable); ok {
|
|
pi := e.table.Meta().GetPartitionInfo()
|
|
for _, p := range pi.Definitions {
|
|
e.table = tbl.GetPartition(p.ID)
|
|
e.index = tables.GetWritableIndexByName(e.index.Meta().Name.L, e.table)
|
|
e.physicalID = p.ID
|
|
addedCnt, scanCnt, err := e.backfillIndex(ctx)
|
|
totalAddedCnt += addedCnt
|
|
totalScanCnt += scanCnt
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
} else {
|
|
totalAddedCnt, totalScanCnt, err = e.backfillIndex(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
req.AppendInt64(0, totalAddedCnt)
|
|
req.AppendInt64(1, totalScanCnt)
|
|
e.done = true
|
|
return nil
|
|
}
|
|
|
|
// CleanupIndexExec represents a cleanup index executor.
|
|
// It is built from "admin cleanup index" statement, is used to delete
|
|
// dangling index data.
|
|
type CleanupIndexExec struct {
|
|
baseExecutor
|
|
|
|
done bool
|
|
removeCnt uint64
|
|
|
|
index table.Index
|
|
table table.Table
|
|
physicalID int64
|
|
|
|
columns []*model.ColumnInfo
|
|
idxColFieldTypes []*types.FieldType
|
|
idxChunk *chunk.Chunk
|
|
handleCols plannercore.HandleCols
|
|
|
|
idxValues *kv.HandleMap // kv.Handle -> [][]types.Datum
|
|
batchSize uint64
|
|
batchKeys []kv.Key
|
|
idxValsBufs [][]types.Datum
|
|
lastIdxKey []byte
|
|
scanRowCnt uint64
|
|
}
|
|
|
|
func (e *CleanupIndexExec) getIdxColTypes() []*types.FieldType {
|
|
if e.idxColFieldTypes != nil {
|
|
return e.idxColFieldTypes
|
|
}
|
|
e.idxColFieldTypes = make([]*types.FieldType, 0, len(e.columns))
|
|
for _, col := range e.columns {
|
|
e.idxColFieldTypes = append(e.idxColFieldTypes, &col.FieldType)
|
|
}
|
|
return e.idxColFieldTypes
|
|
}
|
|
|
|
func (e *CleanupIndexExec) batchGetRecord(txn kv.Transaction) (map[string][]byte, error) {
|
|
e.idxValues.Range(func(h kv.Handle, _ interface{}) bool {
|
|
e.batchKeys = append(e.batchKeys, tablecodec.EncodeRecordKey(e.table.RecordPrefix(), h))
|
|
return true
|
|
})
|
|
values, err := txn.BatchGet(context.Background(), e.batchKeys)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return values, nil
|
|
}
|
|
|
|
func (e *CleanupIndexExec) deleteDanglingIdx(txn kv.Transaction, values map[string][]byte) error {
|
|
for _, k := range e.batchKeys {
|
|
if _, found := values[string(k)]; !found {
|
|
_, handle, err := tablecodec.DecodeRecordKey(k)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
handleIdxValsGroup, ok := e.idxValues.Get(handle)
|
|
if !ok {
|
|
return errors.Trace(errors.Errorf("batch keys are inconsistent with handles"))
|
|
}
|
|
for _, handleIdxVals := range handleIdxValsGroup.([][]types.Datum) {
|
|
if err := e.index.Delete(e.ctx.GetSessionVars().StmtCtx, txn, handleIdxVals, handle); err != nil {
|
|
return err
|
|
}
|
|
e.removeCnt++
|
|
if e.removeCnt%e.batchSize == 0 {
|
|
logutil.BgLogger().Info("clean up dangling index", zap.String("table", e.table.Meta().Name.String()),
|
|
zap.String("index", e.index.Meta().Name.String()), zap.Uint64("count", e.removeCnt))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func extractIdxVals(row chunk.Row, idxVals []types.Datum,
|
|
fieldTypes []*types.FieldType, idxValLen int) []types.Datum {
|
|
if cap(idxVals) < idxValLen {
|
|
idxVals = make([]types.Datum, idxValLen)
|
|
} else {
|
|
idxVals = idxVals[:idxValLen]
|
|
}
|
|
|
|
for i := 0; i < idxValLen; i++ {
|
|
colVal := row.GetDatum(i, fieldTypes[i])
|
|
colVal.Copy(&idxVals[i])
|
|
}
|
|
return idxVals
|
|
}
|
|
|
|
func (e *CleanupIndexExec) fetchIndex(ctx context.Context, txn kv.Transaction) error {
|
|
result, err := e.buildIndexScan(ctx, txn)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer terror.Call(result.Close)
|
|
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
idxColLen := len(e.index.Meta().Columns)
|
|
for {
|
|
err := result.Next(ctx, e.idxChunk)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if e.idxChunk.NumRows() == 0 {
|
|
return nil
|
|
}
|
|
iter := chunk.NewIterator4Chunk(e.idxChunk)
|
|
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
|
|
handle, err := e.handleCols.BuildHandle(row)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
idxVals := extractIdxVals(row, e.idxValsBufs[e.scanRowCnt], e.idxColFieldTypes, idxColLen)
|
|
e.idxValsBufs[e.scanRowCnt] = idxVals
|
|
existingIdxVals, ok := e.idxValues.Get(handle)
|
|
if ok {
|
|
updatedIdxVals := append(existingIdxVals.([][]types.Datum), idxVals)
|
|
e.idxValues.Set(handle, updatedIdxVals)
|
|
} else {
|
|
e.idxValues.Set(handle, [][]types.Datum{idxVals})
|
|
}
|
|
idxKey, _, err := e.index.GenIndexKey(sc, idxVals, handle, nil)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
e.scanRowCnt++
|
|
e.lastIdxKey = idxKey
|
|
if e.scanRowCnt >= e.batchSize {
|
|
return nil
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Next implements the Executor Next interface.
|
|
func (e *CleanupIndexExec) Next(ctx context.Context, req *chunk.Chunk) error {
|
|
req.Reset()
|
|
if e.done {
|
|
return nil
|
|
}
|
|
cleaningClusteredPrimaryKey := e.table.Meta().IsCommonHandle && e.index.Meta().Primary
|
|
if cleaningClusteredPrimaryKey {
|
|
e.done = true
|
|
req.AppendUint64(0, 0)
|
|
return nil
|
|
}
|
|
|
|
var err error
|
|
if tbl, ok := e.table.(table.PartitionedTable); ok {
|
|
pi := e.table.Meta().GetPartitionInfo()
|
|
for _, p := range pi.Definitions {
|
|
e.table = tbl.GetPartition(p.ID)
|
|
e.index = tables.GetWritableIndexByName(e.index.Meta().Name.L, e.table)
|
|
e.physicalID = p.ID
|
|
err = e.init()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = e.cleanTableIndex(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
} else {
|
|
err = e.cleanTableIndex(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
e.done = true
|
|
req.AppendUint64(0, e.removeCnt)
|
|
return nil
|
|
}
|
|
|
|
func (e *CleanupIndexExec) cleanTableIndex(ctx context.Context) error {
|
|
for {
|
|
ctx = kv.WithInternalSourceType(ctx, kv.InternalTxnAdmin)
|
|
errInTxn := kv.RunInNewTxn(ctx, e.ctx.GetStore(), true, func(ctx context.Context, txn kv.Transaction) error {
|
|
txn.SetDiskFullOpt(kvrpcpb.DiskFullOpt_AllowedOnAlmostFull)
|
|
setOptionForTopSQL(e.ctx.GetSessionVars().StmtCtx, txn)
|
|
err := e.fetchIndex(ctx, txn)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
values, err := e.batchGetRecord(txn)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = e.deleteDanglingIdx(txn, values)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
})
|
|
if errInTxn != nil {
|
|
return errInTxn
|
|
}
|
|
if e.scanRowCnt == 0 {
|
|
break
|
|
}
|
|
e.scanRowCnt = 0
|
|
e.batchKeys = e.batchKeys[:0]
|
|
e.idxValues.Range(func(h kv.Handle, val interface{}) bool {
|
|
e.idxValues.Delete(h)
|
|
return true
|
|
})
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (e *CleanupIndexExec) buildIndexScan(ctx context.Context, txn kv.Transaction) (distsql.SelectResult, error) {
|
|
dagPB, err := e.buildIdxDAGPB(txn)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
var builder distsql.RequestBuilder
|
|
ranges := ranger.FullRange()
|
|
kvReq, err := builder.SetIndexRanges(sc, e.physicalID, e.index.Meta().ID, ranges).
|
|
SetDAGRequest(dagPB).
|
|
SetStartTS(txn.StartTS()).
|
|
SetKeepOrder(true).
|
|
SetFromSessionVars(e.ctx.GetSessionVars()).
|
|
SetFromInfoSchema(e.ctx.GetInfoSchema()).
|
|
Build()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
kvReq.KeyRanges[0].StartKey = kv.Key(e.lastIdxKey).PrefixNext()
|
|
kvReq.Concurrency = 1
|
|
result, err := distsql.Select(ctx, e.ctx, kvReq, e.getIdxColTypes(), statistics.NewQueryFeedback(0, nil, 0, false))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return result, nil
|
|
}
|
|
|
|
// Open implements the Executor Open interface.
|
|
func (e *CleanupIndexExec) Open(ctx context.Context) error {
|
|
if err := e.baseExecutor.Open(ctx); err != nil {
|
|
return err
|
|
}
|
|
return e.init()
|
|
}
|
|
|
|
func (e *CleanupIndexExec) init() error {
|
|
e.idxChunk = chunk.New(e.getIdxColTypes(), e.initCap, e.maxChunkSize)
|
|
e.idxValues = kv.NewHandleMap()
|
|
e.batchKeys = make([]kv.Key, 0, e.batchSize)
|
|
e.idxValsBufs = make([][]types.Datum, e.batchSize)
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
idxKey, _, err := e.index.GenIndexKey(sc, []types.Datum{{}}, kv.IntHandle(math.MinInt64), nil)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
e.lastIdxKey = idxKey
|
|
return nil
|
|
}
|
|
|
|
func (e *CleanupIndexExec) buildIdxDAGPB(txn kv.Transaction) (*tipb.DAGRequest, error) {
|
|
dagReq := &tipb.DAGRequest{}
|
|
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
|
|
sc := e.ctx.GetSessionVars().StmtCtx
|
|
dagReq.Flags = sc.PushDownFlags()
|
|
for i := range e.columns {
|
|
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
|
|
}
|
|
|
|
execPB := e.constructIndexScanPB()
|
|
dagReq.Executors = append(dagReq.Executors, execPB)
|
|
err := tables.SetPBColumnsDefaultValue(e.ctx, dagReq.Executors[0].IdxScan.Columns, e.columns)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
limitExec := e.constructLimitPB()
|
|
dagReq.Executors = append(dagReq.Executors, limitExec)
|
|
distsql.SetEncodeType(e.ctx, dagReq)
|
|
return dagReq, nil
|
|
}
|
|
|
|
func (e *CleanupIndexExec) constructIndexScanPB() *tipb.Executor {
|
|
idxExec := &tipb.IndexScan{
|
|
TableId: e.physicalID,
|
|
IndexId: e.index.Meta().ID,
|
|
Columns: util.ColumnsToProto(e.columns, e.table.Meta().PKIsHandle),
|
|
PrimaryColumnIds: tables.TryGetCommonPkColumnIds(e.table.Meta()),
|
|
}
|
|
return &tipb.Executor{Tp: tipb.ExecType_TypeIndexScan, IdxScan: idxExec}
|
|
}
|
|
|
|
func (e *CleanupIndexExec) constructLimitPB() *tipb.Executor {
|
|
limitExec := &tipb.Limit{
|
|
Limit: e.batchSize,
|
|
}
|
|
return &tipb.Executor{Tp: tipb.ExecType_TypeLimit, Limit: limitExec}
|
|
}
|
|
|
|
// Close implements the Executor Close interface.
|
|
func (e *CleanupIndexExec) Close() error {
|
|
return nil
|
|
}
|