181 lines
5.7 KiB
Go
181 lines
5.7 KiB
Go
// Copyright 2020 PingCAP, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package executor
|
|
|
|
import (
|
|
"context"
|
|
|
|
"github.com/pingcap/errors"
|
|
"github.com/pingcap/tidb/distsql"
|
|
"github.com/pingcap/tidb/infoschema"
|
|
"github.com/pingcap/tidb/kv"
|
|
plannercore "github.com/pingcap/tidb/planner/core"
|
|
"github.com/pingcap/tidb/sessionctx"
|
|
"github.com/pingcap/tidb/table"
|
|
"github.com/pingcap/tidb/util/chunk"
|
|
"github.com/pingcap/tidb/util/ranger"
|
|
"github.com/pingcap/tipb/go-tipb"
|
|
)
|
|
|
|
// Currently we only use mpp for broadcast join.
|
|
func useMPPExecution(ctx sessionctx.Context, tr *plannercore.PhysicalTableReader) bool {
|
|
if !ctx.GetSessionVars().AllowMPPExecution {
|
|
return false
|
|
}
|
|
if tr.StoreType != kv.TiFlash {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// MPPGather dispatch MPP tasks and read data from root tasks.
|
|
type MPPGather struct {
|
|
// following fields are construct needed
|
|
baseExecutor
|
|
is infoschema.InfoSchema
|
|
originalPlan plannercore.PhysicalPlan
|
|
startTS uint64
|
|
|
|
allocTaskID int64
|
|
mppReqs []*kv.MPPDispatchRequest
|
|
|
|
respIter distsql.SelectResult
|
|
}
|
|
|
|
func (e *MPPGather) constructMPPTasksImpl(ctx context.Context, p *plannercore.Fragment) ([]*kv.MPPTask, error) {
|
|
if p.TableScan.Table.GetPartitionInfo() == nil {
|
|
return e.constructSinglePhysicalTable(ctx, p.TableScan.Table.ID, p.TableScan.Ranges)
|
|
}
|
|
tmp, _ := e.is.TableByID(p.TableScan.Table.ID)
|
|
tbl := tmp.(table.PartitionedTable)
|
|
partitions, err := partitionPruning(e.ctx, tbl, p.TableScan.PartitionInfo.PruningConds, p.TableScan.PartitionInfo.PartitionNames, p.TableScan.PartitionInfo.Columns, p.TableScan.PartitionInfo.ColumnNames)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
allTasks := make([]*kv.MPPTask, 0)
|
|
for _, part := range partitions {
|
|
partTasks, err := e.constructSinglePhysicalTable(ctx, part.GetPhysicalID(), p.TableScan.Ranges)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
allTasks = append(allTasks, partTasks...)
|
|
}
|
|
return allTasks, nil
|
|
}
|
|
|
|
// single physical table means a table without partitions or a single partition in a partition table.
|
|
func (e *MPPGather) constructSinglePhysicalTable(ctx context.Context, tableID int64, ranges []*ranger.Range) ([]*kv.MPPTask, error) {
|
|
kvRanges := distsql.TableRangesToKVRanges(tableID, ranges, nil)
|
|
req := &kv.MPPBuildTasksRequest{KeyRanges: kvRanges}
|
|
metas, err := e.ctx.GetMPPClient().ConstructMPPTasks(ctx, req)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
tasks := make([]*kv.MPPTask, 0, len(metas))
|
|
for _, meta := range metas {
|
|
e.allocTaskID++
|
|
tasks = append(tasks, &kv.MPPTask{Meta: meta, ID: e.allocTaskID, StartTs: e.startTS, TableID: tableID})
|
|
}
|
|
return tasks, nil
|
|
}
|
|
|
|
func (e *MPPGather) appendMPPDispatchReq(pf *plannercore.Fragment, tasks []*kv.MPPTask, isRoot bool) error {
|
|
dagReq, _, err := constructDAGReq(e.ctx, []plannercore.PhysicalPlan{pf.ExchangeSender}, kv.TiFlash)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
for i := range pf.Schema().Columns {
|
|
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
|
|
}
|
|
if !isRoot {
|
|
dagReq.EncodeType = tipb.EncodeType_TypeCHBlock
|
|
} else {
|
|
dagReq.EncodeType = tipb.EncodeType_TypeChunk
|
|
}
|
|
for _, mppTask := range tasks {
|
|
err := updateExecutorTableID(context.Background(), dagReq.RootExecutor, mppTask.TableID, true)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
pbData, err := dagReq.Marshal()
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
req := &kv.MPPDispatchRequest{
|
|
Data: pbData,
|
|
Meta: mppTask.Meta,
|
|
ID: mppTask.ID,
|
|
IsRoot: isRoot,
|
|
Timeout: 10,
|
|
SchemaVar: e.is.SchemaMetaVersion(),
|
|
StartTs: e.startTS,
|
|
}
|
|
e.mppReqs = append(e.mppReqs, req)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (e *MPPGather) constructMPPTasks(ctx context.Context, pf *plannercore.Fragment, isRoot bool) ([]*kv.MPPTask, error) {
|
|
tasks, err := e.constructMPPTasksImpl(ctx, pf)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
|
|
for _, client := range pf.ExchangeReceivers {
|
|
client.ChildPf.ExchangeSender.Tasks = tasks
|
|
client.Tasks, err = e.constructMPPTasks(ctx, client.ChildPf, false)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
}
|
|
err = e.appendMPPDispatchReq(pf, tasks, isRoot)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
return tasks, nil
|
|
}
|
|
|
|
// Open decides the task counts and locations and generate exchange operators for every plan fragment.
|
|
// Then dispatch tasks to tiflash stores. If any task fails, it would cancel the rest tasks.
|
|
// TODO: We should retry when the request fails for pure rpc error.
|
|
func (e *MPPGather) Open(ctx context.Context) error {
|
|
// TODO: Move the construct tasks logic to planner, so we can see the explain results.
|
|
rootPf := plannercore.GetRootPlanFragments(e.ctx, e.originalPlan, e.startTS)
|
|
_, err := e.constructMPPTasks(ctx, rootPf, true)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
|
|
e.respIter, err = distsql.DispatchMPPTasks(ctx, e.ctx, e.mppReqs, e.retFieldTypes)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
e.respIter.Fetch(ctx)
|
|
return nil
|
|
}
|
|
|
|
// Next fills data into the chunk passed by its caller.
|
|
func (e *MPPGather) Next(ctx context.Context, chk *chunk.Chunk) error {
|
|
err := e.respIter.Next(ctx, chk)
|
|
return errors.Trace(err)
|
|
}
|
|
|
|
// Close and release the used resources.
|
|
func (e *MPPGather) Close() error {
|
|
if e.respIter != nil {
|
|
return e.respIter.Close()
|
|
}
|
|
return nil
|
|
}
|