200 lines
5.5 KiB
Go
200 lines
5.5 KiB
Go
// Copyright 2018 PingCAP, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package executor
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"time"
|
|
|
|
"github.com/opentracing/basictracer-go"
|
|
"github.com/opentracing/opentracing-go"
|
|
"github.com/pingcap/errors"
|
|
"github.com/pingcap/parser/ast"
|
|
"github.com/pingcap/tidb/planner"
|
|
plannercore "github.com/pingcap/tidb/planner/core"
|
|
"github.com/pingcap/tidb/sessionctx"
|
|
"github.com/pingcap/tidb/util/chunk"
|
|
"github.com/pingcap/tidb/util/sqlexec"
|
|
"github.com/pingcap/tidb/util/tracing"
|
|
"sourcegraph.com/sourcegraph/appdash"
|
|
traceImpl "sourcegraph.com/sourcegraph/appdash/opentracing"
|
|
)
|
|
|
|
// TraceExec represents a root executor of trace query.
|
|
type TraceExec struct {
|
|
baseExecutor
|
|
// CollectedSpans collects all span during execution. Span is appended via
|
|
// callback method which passes into tracer implementation.
|
|
CollectedSpans []basictracer.RawSpan
|
|
// exhausted being true means there is no more result.
|
|
exhausted bool
|
|
// stmtNode is the real query ast tree and it is used for building real query's plan.
|
|
stmtNode ast.StmtNode
|
|
// rootTrace represents root span which is father of all other span.
|
|
rootTrace opentracing.Span
|
|
|
|
builder *executorBuilder
|
|
format string
|
|
}
|
|
|
|
// Next executes real query and collects span later.
|
|
func (e *TraceExec) Next(ctx context.Context, chk *chunk.Chunk) error {
|
|
chk.Reset()
|
|
if e.exhausted {
|
|
return nil
|
|
}
|
|
|
|
if e.format == "json" {
|
|
if se, ok := e.ctx.(sqlexec.SQLExecutor); ok {
|
|
store := appdash.NewMemoryStore()
|
|
tracer := traceImpl.NewTracer(store)
|
|
span := tracer.StartSpan("trace")
|
|
defer span.Finish()
|
|
ctx = opentracing.ContextWithSpan(ctx, span)
|
|
recordSets, err := se.Execute(ctx, e.stmtNode.Text())
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
|
|
for _, rs := range recordSets {
|
|
_, err = drainRecordSet(ctx, e.ctx, rs)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
if err = rs.Close(); err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
}
|
|
|
|
traces, err := store.Traces(appdash.TracesOpts{})
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
data, err := json.Marshal(traces)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
|
|
// Split json data into rows to avoid the max packet size limitation.
|
|
const maxRowLen = 4096
|
|
for len(data) > maxRowLen {
|
|
chk.AppendString(0, string(data[:maxRowLen]))
|
|
data = data[maxRowLen:]
|
|
}
|
|
chk.AppendString(0, string(data))
|
|
}
|
|
e.exhausted = true
|
|
return nil
|
|
}
|
|
|
|
// TODO: If the following code is never used, remove it later.
|
|
// record how much time was spent for optimizeing plan
|
|
optimizeSp := e.rootTrace.Tracer().StartSpan("plan_optimize", opentracing.FollowsFrom(e.rootTrace.Context()))
|
|
stmtPlan, err := planner.Optimize(e.builder.ctx, e.stmtNode, e.builder.is)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
optimizeSp.Finish()
|
|
|
|
pp, ok := stmtPlan.(plannercore.PhysicalPlan)
|
|
if !ok {
|
|
return errors.New("cannot cast logical plan to physical plan")
|
|
}
|
|
|
|
// append select executor to trace executor
|
|
stmtExec := e.builder.build(pp)
|
|
|
|
e.rootTrace = tracing.NewRecordedTrace("trace_exec", func(sp basictracer.RawSpan) {
|
|
e.CollectedSpans = append(e.CollectedSpans, sp)
|
|
})
|
|
err = stmtExec.Open(ctx)
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
stmtExecChk := stmtExec.newFirstChunk()
|
|
|
|
// store span into context
|
|
ctx = opentracing.ContextWithSpan(ctx, e.rootTrace)
|
|
|
|
for {
|
|
if err := stmtExec.Next(ctx, stmtExecChk); err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
if stmtExecChk.NumRows() == 0 {
|
|
break
|
|
}
|
|
}
|
|
|
|
e.rootTrace.LogKV("event", "tracing completed")
|
|
e.rootTrace.Finish()
|
|
var rootSpan basictracer.RawSpan
|
|
|
|
treeSpans := make(map[uint64][]basictracer.RawSpan)
|
|
for _, sp := range e.CollectedSpans {
|
|
treeSpans[sp.ParentSpanID] = append(treeSpans[sp.ParentSpanID], sp)
|
|
// if a span's parentSpanID is 0, then it is root span
|
|
// this is by design
|
|
if sp.ParentSpanID == 0 {
|
|
rootSpan = sp
|
|
}
|
|
}
|
|
|
|
dfsTree(rootSpan, treeSpans, "", false, chk)
|
|
e.exhausted = true
|
|
return nil
|
|
}
|
|
|
|
func drainRecordSet(ctx context.Context, sctx sessionctx.Context, rs sqlexec.RecordSet) ([]chunk.Row, error) {
|
|
var rows []chunk.Row
|
|
chk := rs.NewChunk()
|
|
|
|
for {
|
|
err := rs.Next(ctx, chk)
|
|
if err != nil || chk.NumRows() == 0 {
|
|
return rows, errors.Trace(err)
|
|
}
|
|
iter := chunk.NewIterator4Chunk(chk)
|
|
for r := iter.Begin(); r != iter.End(); r = iter.Next() {
|
|
rows = append(rows, r)
|
|
}
|
|
chk = chunk.Renew(chk, sctx.GetSessionVars().MaxChunkSize)
|
|
}
|
|
}
|
|
|
|
func dfsTree(span basictracer.RawSpan, tree map[uint64][]basictracer.RawSpan, prefix string, isLast bool, chk *chunk.Chunk) {
|
|
suffix := ""
|
|
spans := tree[span.Context.SpanID]
|
|
var newPrefix string
|
|
if span.ParentSpanID == 0 {
|
|
newPrefix = prefix
|
|
} else {
|
|
if len(tree[span.ParentSpanID]) > 0 && !isLast {
|
|
suffix = "├─"
|
|
newPrefix = prefix + "│ "
|
|
} else {
|
|
suffix = "└─"
|
|
newPrefix = prefix + " "
|
|
}
|
|
}
|
|
|
|
chk.AppendString(0, prefix+suffix+span.Operation)
|
|
chk.AppendString(1, span.Start.Format(time.StampNano))
|
|
chk.AppendString(2, span.Duration.String())
|
|
|
|
for i, sp := range spans {
|
|
dfsTree(sp, tree, newPrefix, i == (len(spans))-1 /*last element of array*/, chk)
|
|
}
|
|
}
|