// Copyright 2016 PingCAP, Inc. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package executor import ( "runtime" "strconv" "testing" "time" "unsafe" "github.com/pingcap/tidb/pkg/executor/aggfuncs" "github.com/pingcap/tidb/pkg/executor/aggregate" "github.com/pingcap/tidb/pkg/kv" "github.com/pingcap/tidb/pkg/sessionctx/variable" "github.com/pingcap/tidb/pkg/tablecodec" "github.com/pingcap/tidb/pkg/types" "github.com/pingcap/tidb/pkg/util/collate" "github.com/pingcap/tidb/pkg/util/memory" "github.com/pingcap/tidb/pkg/util/mock" "github.com/pingcap/tidb/pkg/util/ranger" "github.com/pingcap/tidb/pkg/util/tableutil" "github.com/stretchr/testify/require" ) // Note: it's a tricky way to export the `inspectionSummaryRules` and `inspectionRules` for unit test but invisible for normal code var ( InspectionSummaryRules = inspectionSummaryRules InspectionRules = inspectionRules ) func TestBuildKvRangesForIndexJoinWithoutCwc(t *testing.T) { indexRanges := make([]*ranger.Range, 0, 6) indexRanges = append(indexRanges, generateIndexRange(1, 1, 1, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 2, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 2, 1, 2)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 3, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(2, 1, 1, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(2, 1, 2, 1, 1)) joinKeyRows := make([]*indexJoinLookUpContent, 0, 5) joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(1, 1)}) joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(1, 2)}) joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(2, 1)}) joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(2, 2)}) joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(2, 3)}) keyOff2IdxOff := []int{1, 3} ctx := mock.NewContext() kvRanges, err := buildKvRangesForIndexJoin(ctx, 0, 0, joinKeyRows, indexRanges, keyOff2IdxOff, nil, nil, nil) require.NoError(t, err) // Check the kvRanges is in order. for i, kvRange := range kvRanges { require.True(t, kvRange.StartKey.Cmp(kvRange.EndKey) < 0) if i > 0 { require.True(t, kvRange.StartKey.Cmp(kvRanges[i-1].EndKey) >= 0) } } } func TestBuildKvRangesForIndexJoinWithoutCwcAndWithMemoryTracker(t *testing.T) { indexRanges := make([]*ranger.Range, 0, 6) indexRanges = append(indexRanges, generateIndexRange(1, 1, 1, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 2, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 2, 1, 2)) indexRanges = append(indexRanges, generateIndexRange(1, 1, 3, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(2, 1, 1, 1, 1)) indexRanges = append(indexRanges, generateIndexRange(2, 1, 2, 1, 1)) bytesConsumed1 := int64(0) { joinKeyRows := make([]*indexJoinLookUpContent, 0, 10) for i := int64(0); i < 10; i++ { joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(1, i)}) } keyOff2IdxOff := []int{1, 3} ctx := mock.NewContext() memTracker := memory.NewTracker(memory.LabelForIndexWorker, -1) kvRanges, err := buildKvRangesForIndexJoin(ctx, 0, 0, joinKeyRows, indexRanges, keyOff2IdxOff, nil, memTracker, nil) require.NoError(t, err) // Check the kvRanges is in order. for i, kvRange := range kvRanges { require.True(t, kvRange.StartKey.Cmp(kvRange.EndKey) < 0) if i > 0 { require.True(t, kvRange.StartKey.Cmp(kvRanges[i-1].EndKey) >= 0) } } bytesConsumed1 = memTracker.BytesConsumed() } bytesConsumed2 := int64(0) { joinKeyRows := make([]*indexJoinLookUpContent, 0, 20) for i := int64(0); i < 20; i++ { joinKeyRows = append(joinKeyRows, &indexJoinLookUpContent{keys: generateDatumSlice(1, i)}) } keyOff2IdxOff := []int{1, 3} ctx := mock.NewContext() memTracker := memory.NewTracker(memory.LabelForIndexWorker, -1) kvRanges, err := buildKvRangesForIndexJoin(ctx, 0, 0, joinKeyRows, indexRanges, keyOff2IdxOff, nil, memTracker, nil) require.NoError(t, err) // Check the kvRanges is in order. for i, kvRange := range kvRanges { require.True(t, kvRange.StartKey.Cmp(kvRange.EndKey) < 0) if i > 0 { require.True(t, kvRange.StartKey.Cmp(kvRanges[i-1].EndKey) >= 0) } } bytesConsumed2 = memTracker.BytesConsumed() } require.Equal(t, 2*bytesConsumed1, bytesConsumed2) require.Equal(t, int64(20760), bytesConsumed1) } func generateIndexRange(vals ...int64) *ranger.Range { lowDatums := generateDatumSlice(vals...) highDatums := make([]types.Datum, len(vals)) copy(highDatums, lowDatums) return &ranger.Range{LowVal: lowDatums, HighVal: highDatums, Collators: collate.GetBinaryCollatorSlice(len(lowDatums))} } func generateDatumSlice(vals ...int64) []types.Datum { datums := make([]types.Datum, len(vals)) for i, val := range vals { datums[i].SetInt64(val) } return datums } func TestSlowQueryRuntimeStats(t *testing.T) { stats := &slowQueryRuntimeStats{ totalFileNum: 2, readFileNum: 2, readFile: time.Second, initialize: time.Millisecond, readFileSize: 1024 * 1024 * 1024, parseLog: int64(time.Millisecond * 100), concurrent: 15, } require.Equal(t, "initialize: 1ms, read_file: 1s, parse_log: {time:100ms, concurrency:15}, total_file: 2, read_file: 2, read_size: 1024 MB", stats.String()) require.Equal(t, stats.Clone().String(), stats.String()) stats.Merge(stats.Clone()) require.Equal(t, "initialize: 2ms, read_file: 2s, parse_log: {time:200ms, concurrency:15}, total_file: 4, read_file: 4, read_size: 2 GB", stats.String()) } // Test whether the actual buckets in Golang Map is same with the estimated number. // The test relies the implement of Golang Map. ref https://github.com/golang/go/blob/go1.13/src/runtime/map.go#L114 func TestAggPartialResultMapperB(t *testing.T) { if runtime.Version() < `go1.13` { t.Skip("Unsupported version") } type testCase struct { rowNum int expectedB int expectedGrowing bool } cases := []testCase{ { rowNum: 0, expectedB: 0, expectedGrowing: false, }, { rowNum: 100, expectedB: 5, expectedGrowing: true, }, { rowNum: 10000, expectedB: 11, expectedGrowing: false, }, { rowNum: 1000000, expectedB: 18, expectedGrowing: false, }, { rowNum: 851968, // 6.5 * (1 << 17) expectedB: 18, expectedGrowing: true, }, { rowNum: 851969, // 6.5 * (1 << 17) + 1 expectedB: 18, expectedGrowing: true, }, { rowNum: 425984, // 6.5 * (1 << 16) expectedB: 17, expectedGrowing: true, }, { rowNum: 425985, // 6.5 * (1 << 16) + 1 expectedB: 17, expectedGrowing: true, }, } for _, tc := range cases { aggMap := make(aggregate.AggPartialResultMapper) tempSlice := make([]aggfuncs.PartialResult, 10) for num := 0; num < tc.rowNum; num++ { aggMap[strconv.Itoa(num)] = tempSlice } require.Equal(t, tc.expectedB, getB(aggMap)) require.Equal(t, tc.expectedGrowing, getGrowing(aggMap)) } } // A header for a Go map. // nolint:structcheck type hmap struct { // Note: the format of the hmap is also encoded in cmd/compile/internal/gc/reflect.go. // Make sure this stays in sync with the compiler's definition. count int // nolint:unused // # live cells == size of map. Must be first (used by len() builtin) flags uint8 // nolint:unused B uint8 // nolint:unused // log_2 of # of buckets (can hold up to loadFactor * 2^B items) noverflow uint16 // nolint:unused // approximate number of overflow buckets; see incrnoverflow for details hash0 uint32 // nolint:unused // hash seed buckets unsafe.Pointer // nolint:unused // array of 2^B Buckets. may be nil if count==0. oldbuckets unsafe.Pointer // nolint:unused // previous bucket array of half the size, non-nil only when growing nevacuate uintptr // nolint:unused // progress counter for evacuation (buckets less than this have been evacuated) } func getB(m aggregate.AggPartialResultMapper) int { point := (**hmap)(unsafe.Pointer(&m)) value := *point return int(value.B) } func getGrowing(m aggregate.AggPartialResultMapper) bool { point := (**hmap)(unsafe.Pointer(&m)) value := *point return value.oldbuckets != nil } func TestFilterTemporaryTableKeys(t *testing.T) { vars := variable.NewSessionVars(nil) const tableID int64 = 3 vars.TxnCtx = &variable.TransactionContext{ TxnCtxNoNeedToRestore: variable.TxnCtxNoNeedToRestore{ TemporaryTables: map[int64]tableutil.TempTable{tableID: nil}, }, } res := filterTemporaryTableKeys(vars, []kv.Key{tablecodec.EncodeTablePrefix(tableID), tablecodec.EncodeTablePrefix(42)}) require.Len(t, res, 1) }