Files
tidb/executor/importer/engine_process.go
2023-06-09 07:58:47 +00:00

107 lines
3.4 KiB
Go

// Copyright 2023 PingCAP, Inc.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package importer
import (
"context"
"github.com/pingcap/tidb/br/pkg/lightning/backend"
"github.com/pingcap/tidb/br/pkg/lightning/checkpoints"
"github.com/pingcap/tidb/br/pkg/lightning/common"
"github.com/pingcap/tidb/executor/asyncloaddata"
"go.uber.org/zap"
)
// ProcessChunk processes a chunk, and write kv pairs to dataEngine and indexEngine.
func ProcessChunk(
ctx context.Context,
chunk *checkpoints.ChunkCheckpoint,
tableImporter *TableImporter,
dataEngine,
indexEngine *backend.OpenedEngine,
progress *asyncloaddata.Progress,
logger *zap.Logger,
) error {
// if the key are ordered, LocalWrite can optimize the writing.
// table has auto-incremented _tidb_rowid must satisfy following restrictions:
// - clustered index disable and primary key is not number
// - no auto random bits (auto random or shard row id)
// - no partition table
// - no explicit _tidb_rowid field (At this time we can't determine if the source file contains _tidb_rowid field,
// so we will do this check in LocalWriter when the first row is received.)
hasAutoIncrementAutoID := common.TableHasAutoRowID(tableImporter.tableInfo.Core) &&
tableImporter.tableInfo.Core.AutoRandomBits == 0 && tableImporter.tableInfo.Core.ShardRowIDBits == 0 &&
tableImporter.tableInfo.Core.Partition == nil
dataWriterCfg := &backend.LocalWriterConfig{
IsKVSorted: hasAutoIncrementAutoID,
}
parser, err := tableImporter.getParser(ctx, chunk)
if err != nil {
return err
}
defer func() {
if err2 := parser.Close(); err2 != nil {
logger.Warn("close parser failed", zap.Error(err2))
}
}()
encoder, err := tableImporter.getKVEncoder(chunk)
if err != nil {
return err
}
defer func() {
if err2 := encoder.Close(); err2 != nil {
logger.Warn("close encoder failed", zap.Error(err2))
}
}()
dataWriter, err := dataEngine.LocalWriter(ctx, dataWriterCfg)
if err != nil {
return err
}
defer func() {
if _, err2 := dataWriter.Close(ctx); err2 != nil {
logger.Warn("close data writer failed", zap.Error(err2))
}
}()
indexWriter, err := indexEngine.LocalWriter(ctx, &backend.LocalWriterConfig{})
if err != nil {
return err
}
defer func() {
if _, err2 := indexWriter.Close(ctx); err2 != nil {
logger.Warn("close index writer failed", zap.Error(err2))
}
}()
cp := &chunkProcessor{
parser: parser,
chunkInfo: chunk,
logger: logger.With(zap.String("key", chunk.GetKey())),
kvsCh: make(chan []deliveredRow, maxKVQueueSize),
dataWriter: dataWriter,
indexWriter: indexWriter,
encoder: encoder,
kvCodec: tableImporter.kvStore.GetCodec(),
progress: progress,
diskQuotaLock: tableImporter.diskQuotaLock,
}
// todo: process in parallel
err = cp.process(ctx)
if err != nil {
return err
}
tableImporter.setLastInsertID(encoder.GetLastInsertID())
return nil
}