Files
tidb/pkg/util/execdetails/execdetails_test.go

493 lines
22 KiB
Go

// Copyright 2018 PingCAP, Inc.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package execdetails
import (
"fmt"
"strconv"
"sync"
"testing"
"time"
"github.com/pingcap/tipb/go-tipb"
"github.com/stretchr/testify/require"
"github.com/tikv/client-go/v2/util"
)
func TestString(t *testing.T) {
detail := &ExecDetails{
CopTime: time.Second + 3*time.Millisecond,
BackoffTime: time.Second,
RequestCount: 1,
CommitDetail: &util.CommitDetails{
GetCommitTsTime: time.Second,
GetLatestTsTime: time.Second,
PrewriteTime: time.Second,
CommitTime: time.Second,
LocalLatchTime: time.Second,
Mu: struct {
sync.Mutex
CommitBackoffTime int64
PrewriteBackoffTypes []string
CommitBackoffTypes []string
SlowestPrewrite util.ReqDetailInfo
CommitPrimary util.ReqDetailInfo
}{
CommitBackoffTime: int64(time.Second),
PrewriteBackoffTypes: []string{
"backoff1",
"backoff2",
},
CommitBackoffTypes: []string{
"commit1",
"commit2",
},
SlowestPrewrite: util.ReqDetailInfo{
ReqTotalTime: time.Second,
Region: 1000,
StoreAddr: "tikv-1:20160",
ExecDetails: util.TiKVExecDetails{
TimeDetail: &util.TimeDetail{
TotalRPCWallTime: 500 * time.Millisecond,
},
ScanDetail: &util.ScanDetail{
ProcessedKeys: 10,
TotalKeys: 100,
RocksdbDeleteSkippedCount: 1,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 1,
RocksdbBlockReadCount: 1,
RocksdbBlockReadByte: 100,
RocksdbBlockReadDuration: 20 * time.Millisecond,
},
WriteDetail: &util.WriteDetail{
StoreBatchWaitDuration: 10 * time.Microsecond,
ProposeSendWaitDuration: 20 * time.Microsecond,
PersistLogDuration: 30 * time.Microsecond,
RaftDbWriteLeaderWaitDuration: 40 * time.Microsecond,
RaftDbSyncLogDuration: 45 * time.Microsecond,
RaftDbWriteMemtableDuration: 50 * time.Microsecond,
CommitLogDuration: 60 * time.Microsecond,
ApplyBatchWaitDuration: 70 * time.Microsecond,
ApplyLogDuration: 80 * time.Microsecond,
ApplyMutexLockDuration: 90 * time.Microsecond,
ApplyWriteLeaderWaitDuration: 100 * time.Microsecond,
ApplyWriteWalDuration: 101 * time.Microsecond,
ApplyWriteMemtableDuration: 102 * time.Microsecond,
},
},
},
CommitPrimary: util.ReqDetailInfo{},
},
WriteKeys: 1,
WriteSize: 1,
PrewriteRegionNum: 1,
TxnRetry: 1,
ResolveLock: util.ResolveLockDetail{
ResolveLockTime: 1000000000, // 10^9 ns = 1s
},
},
ScanDetail: &util.ScanDetail{
ProcessedKeys: 10,
TotalKeys: 100,
RocksdbDeleteSkippedCount: 1,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 1,
RocksdbBlockReadCount: 1,
RocksdbBlockReadByte: 100,
RocksdbBlockReadDuration: time.Millisecond,
},
DetailsNeedP90: DetailsNeedP90{TimeDetail: util.TimeDetail{
ProcessTime: 2*time.Second + 5*time.Millisecond,
WaitTime: time.Second,
}},
}
expected := "Cop_time: 1.003 Process_time: 2.005 Wait_time: 1 Backoff_time: 1 Request_count: 1 Prewrite_time: 1 Commit_time: " +
"1 Get_commit_ts_time: 1 Get_latest_ts_time: 1 Commit_backoff_time: 1 " +
"Prewrite_Backoff_types: [backoff1 backoff2] Commit_Backoff_types: [commit1 commit2] Slowest_prewrite_rpc_detail: {total:1.000s, region_id: 1000, " +
"store: tikv-1:20160, tikv_wall_time: 500ms, scan_detail: {total_process_keys: 10, total_keys: 100, " +
"rocksdb: {delete_skipped_count: 1, key_skipped_count: 1, block: {cache_hit_count: 1, read_count: 1, " +
"read_byte: 100 Bytes, read_time: 20ms}}}, write_detail: {store_batch_wait: 10µs, propose_send_wait: 20µs, " +
"persist_log: {total: 30µs, write_leader_wait: 40µs, sync_log: 45µs, write_memtable: 50µs}, " +
"commit_log: 60µs, apply_batch_wait: 70µs, apply: {total:80µs, mutex_lock: 90µs, write_leader_wait: 100µs, " +
"write_wal: 101µs, write_memtable: 102µs}}} Resolve_lock_time: 1 Local_latch_wait_time: 1 Write_keys: 1 Write_size: " +
"1 Prewrite_region: 1 Txn_retry: 1 Process_keys: 10 Total_keys: 100 Rocksdb_delete_skipped_count: 1 Rocksdb_key_skipped_count: " +
"1 Rocksdb_block_cache_hit_count: 1 Rocksdb_block_read_count: 1 Rocksdb_block_read_byte: 100 Rocksdb_block_read_time: 0.001"
require.Equal(t, expected, detail.String())
detail = &ExecDetails{}
require.Equal(t, "", detail.String())
}
func mockExecutorExecutionSummary(TimeProcessedNs, NumProducedRows, NumIterations uint64) *tipb.ExecutorExecutionSummary {
return &tipb.ExecutorExecutionSummary{TimeProcessedNs: &TimeProcessedNs, NumProducedRows: &NumProducedRows,
NumIterations: &NumIterations, XXX_unrecognized: nil}
}
func mockExecutorExecutionSummaryForTiFlash(TimeProcessedNs, NumProducedRows, NumIterations, Concurrency, dmfileScannedRows, dmfileSkippedRows, totalDmfileRsCheckMs, totalDmfileReadTimeMs, totalBuildSnapshotMs, localRegions, remoteRegions, totalLearnerReadMs, disaggReadCacheHitBytes, disaggReadCacheMissBytes uint64, ExecutorID string) *tipb.ExecutorExecutionSummary {
tiflashScanContext := tipb.TiFlashScanContext{
DmfileDataScannedRows: &dmfileScannedRows,
DmfileDataSkippedRows: &dmfileSkippedRows,
TotalDmfileRsCheckMs: &totalDmfileRsCheckMs,
TotalDmfileReadMs: &totalDmfileReadTimeMs,
TotalBuildSnapshotMs: &totalBuildSnapshotMs,
LocalRegions: &localRegions,
RemoteRegions: &remoteRegions,
TotalLearnerReadMs: &totalLearnerReadMs,
DisaggReadCacheHitBytes: &disaggReadCacheHitBytes,
DisaggReadCacheMissBytes: &disaggReadCacheMissBytes,
}
fmt.Println("tiflashScanContext is ", tiflashScanContext.String())
return &tipb.ExecutorExecutionSummary{TimeProcessedNs: &TimeProcessedNs, NumProducedRows: &NumProducedRows,
NumIterations: &NumIterations, Concurrency: &Concurrency, ExecutorId: &ExecutorID, DetailInfo: &tipb.ExecutorExecutionSummary_TiflashScanContext{TiflashScanContext: &tiflashScanContext}, XXX_unrecognized: nil}
}
func TestCopRuntimeStats(t *testing.T) {
stats := NewRuntimeStatsColl(nil)
tableScanID := 1
aggID := 2
tableReaderID := 3
stats.RecordOneCopTask(tableScanID, "tikv", "8.8.8.8", mockExecutorExecutionSummary(1, 1, 1))
stats.RecordOneCopTask(tableScanID, "tikv", "8.8.8.9", mockExecutorExecutionSummary(2, 2, 2))
stats.RecordOneCopTask(aggID, "tikv", "8.8.8.8", mockExecutorExecutionSummary(3, 3, 3))
stats.RecordOneCopTask(aggID, "tikv", "8.8.8.9", mockExecutorExecutionSummary(4, 4, 4))
scanDetail := &util.ScanDetail{
TotalKeys: 15,
ProcessedKeys: 10,
ProcessedKeysSize: 10,
RocksdbDeleteSkippedCount: 5,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 10,
RocksdbBlockReadCount: 20,
RocksdbBlockReadByte: 100,
}
stats.RecordScanDetail(tableScanID, "tikv", scanDetail)
require.True(t, stats.ExistsCopStats(tableScanID))
cop := stats.GetOrCreateCopStats(tableScanID, "tikv")
expected := "tikv_task:{proc max:2ns, min:1ns, avg: 1ns, p80:2ns, p95:2ns, iters:3, tasks:2}, " +
"scan_detail: {total_process_keys: 10, total_process_keys_size: 10, total_keys: 15, rocksdb: {delete_skipped_count: 5, key_skipped_count: 1, block: {cache_hit_count: 10, read_count: 20, read_byte: 100 Bytes}}}"
require.Equal(t, expected, cop.String())
copStats := cop.stats["8.8.8.8"]
require.NotNil(t, copStats)
newCopStats := &basicCopRuntimeStats{}
newCopStats.SetRowNum(10)
newCopStats.Record(time.Second, 10)
copStats.Merge(newCopStats)
require.Equal(t, "time:1s, loops:2", copStats.String())
require.Equal(t, "tikv_task:{proc max:4ns, min:3ns, avg: 3ns, p80:4ns, p95:4ns, iters:7, tasks:2}", stats.GetOrCreateCopStats(aggID, "tikv").String())
rootStats := stats.GetRootStats(tableReaderID)
require.NotNil(t, rootStats)
require.True(t, stats.ExistsRootStats(tableReaderID))
cop.scanDetail.ProcessedKeys = 0
cop.scanDetail.ProcessedKeysSize = 0
cop.scanDetail.RocksdbKeySkippedCount = 0
cop.scanDetail.RocksdbBlockReadCount = 0
// Print all fields even though the value of some fields is 0.
str := "tikv_task:{proc max:1s, min:1ns, avg: 500ms, p80:1s, p95:1s, iters:4, tasks:2}, " +
"scan_detail: {total_keys: 15, rocksdb: {delete_skipped_count: 5, block: {cache_hit_count: 10, read_byte: 100 Bytes}}}"
require.Equal(t, str, cop.String())
zeroScanDetail := util.ScanDetail{}
require.Equal(t, "", zeroScanDetail.String())
}
func TestCopRuntimeStatsForTiFlash(t *testing.T) {
stats := NewRuntimeStatsColl(nil)
tableScanID := 1
aggID := 2
tableReaderID := 3
stats.RecordOneCopTask(aggID, "tiflash", "8.8.8.8", mockExecutorExecutionSummaryForTiFlash(1, 1, 1, 1, 8192, 0, 15, 200, 40, 10, 4, 1, 100, 50, "tablescan_"+strconv.Itoa(tableScanID)))
stats.RecordOneCopTask(aggID, "tiflash", "8.8.8.9", mockExecutorExecutionSummaryForTiFlash(2, 2, 2, 1, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, "tablescan_"+strconv.Itoa(tableScanID)))
stats.RecordOneCopTask(tableScanID, "tiflash", "8.8.8.8", mockExecutorExecutionSummaryForTiFlash(3, 3, 3, 1, 12000, 6000, 60, 1000, 20, 5, 1, 0, 20, 0, "aggregation_"+strconv.Itoa(aggID)))
stats.RecordOneCopTask(tableScanID, "tiflash", "8.8.8.9", mockExecutorExecutionSummaryForTiFlash(4, 4, 4, 1, 8192, 80000, 40, 2000, 30, 1, 1, 0, 0, 0, "aggregation_"+strconv.Itoa(aggID)))
scanDetail := &util.ScanDetail{
TotalKeys: 10,
ProcessedKeys: 10,
RocksdbDeleteSkippedCount: 10,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 10,
RocksdbBlockReadCount: 10,
RocksdbBlockReadByte: 100,
}
stats.RecordScanDetail(tableScanID, "tiflash", scanDetail)
require.True(t, stats.ExistsCopStats(tableScanID))
cop := stats.GetOrCreateCopStats(tableScanID, "tiflash")
require.Equal(t, "tiflash_task:{proc max:2ns, min:1ns, avg: 1ns, p80:2ns, p95:2ns, iters:3, tasks:2, threads:2}, tiflash_scan:{mvcc_input_rows:0, mvcc_input_bytes:0, mvcc_output_rows:0, lm_skip_rows:0, local_regions:10, remote_regions:4, tot_learner_read:1ms, region_balance:none, delta_rows:0, delta_bytes:0, segments:0, stale_read_regions:0, tot_build_snapshot:40ms, tot_build_bitmap:0ms, tot_build_inputstream:0ms, min_local_stream:0ms, max_local_stream:0ms, dtfile:{data_scanned_rows:8192, data_skipped_rows:0, mvcc_scanned_rows:0, mvcc_skipped_rows:0, lm_filter_scanned_rows:0, lm_filter_skipped_rows:0, tot_rs_index_check:15ms, tot_read:202ms, disagg_cache_hit_bytes: 100, disagg_cache_miss_bytes: 50}}", cop.String())
copStats := cop.stats["8.8.8.8"]
require.NotNil(t, copStats)
copStats.SetRowNum(10)
copStats.Record(time.Second, 10)
require.Equal(t, "time:1s, loops:2, threads:1, tiflash_scan:{mvcc_input_rows:0, mvcc_input_bytes:0, mvcc_output_rows:0, lm_skip_rows:0, local_regions:10, remote_regions:4, tot_learner_read:1ms, region_balance:none, delta_rows:0, delta_bytes:0, segments:0, stale_read_regions:0, tot_build_snapshot:40ms, tot_build_bitmap:0ms, tot_build_inputstream:0ms, min_local_stream:0ms, max_local_stream:0ms, dtfile:{data_scanned_rows:8192, data_skipped_rows:0, mvcc_scanned_rows:0, mvcc_skipped_rows:0, lm_filter_scanned_rows:0, lm_filter_skipped_rows:0, tot_rs_index_check:15ms, tot_read:200ms, disagg_cache_hit_bytes: 100, disagg_cache_miss_bytes: 50}}", copStats.String())
expected := "tiflash_task:{proc max:4ns, min:3ns, avg: 3ns, p80:4ns, p95:4ns, iters:7, tasks:2, threads:2}, tiflash_scan:{mvcc_input_rows:0, mvcc_input_bytes:0, mvcc_output_rows:0, lm_skip_rows:0, local_regions:6, remote_regions:2, tot_learner_read:0ms, region_balance:none, delta_rows:0, delta_bytes:0, segments:0, stale_read_regions:0, tot_build_snapshot:50ms, tot_build_bitmap:0ms, tot_build_inputstream:0ms, min_local_stream:0ms, max_local_stream:0ms, dtfile:{data_scanned_rows:20192, data_skipped_rows:86000, mvcc_scanned_rows:0, mvcc_skipped_rows:0, lm_filter_scanned_rows:0, lm_filter_skipped_rows:0, tot_rs_index_check:100ms, tot_read:3000ms, disagg_cache_hit_bytes: 20, disagg_cache_miss_bytes: 0}}"
require.Equal(t, expected, stats.GetOrCreateCopStats(aggID, "tiflash").String())
rootStats := stats.GetRootStats(tableReaderID)
require.NotNil(t, rootStats)
require.True(t, stats.ExistsRootStats(tableReaderID))
}
func TestRuntimeStatsWithCommit(t *testing.T) {
commitDetail := &util.CommitDetails{
GetCommitTsTime: time.Second,
PrewriteTime: time.Second,
CommitTime: time.Second,
Mu: struct {
sync.Mutex
CommitBackoffTime int64
PrewriteBackoffTypes []string
CommitBackoffTypes []string
SlowestPrewrite util.ReqDetailInfo
CommitPrimary util.ReqDetailInfo
}{
CommitBackoffTime: int64(time.Second),
PrewriteBackoffTypes: []string{"backoff1", "backoff2", "backoff1"},
CommitBackoffTypes: []string{},
SlowestPrewrite: util.ReqDetailInfo{
ReqTotalTime: time.Second,
Region: 1000,
StoreAddr: "tikv-1:20160",
ExecDetails: util.TiKVExecDetails{
TimeDetail: &util.TimeDetail{
TotalRPCWallTime: 500 * time.Millisecond,
},
ScanDetail: &util.ScanDetail{
ProcessedKeys: 10,
TotalKeys: 100,
RocksdbDeleteSkippedCount: 1,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 1,
RocksdbBlockReadCount: 1,
RocksdbBlockReadByte: 100,
RocksdbBlockReadDuration: 20 * time.Millisecond,
},
WriteDetail: &util.WriteDetail{
StoreBatchWaitDuration: 10 * time.Microsecond,
ProposeSendWaitDuration: 20 * time.Microsecond,
PersistLogDuration: 30 * time.Microsecond,
RaftDbWriteLeaderWaitDuration: 40 * time.Microsecond,
RaftDbSyncLogDuration: 45 * time.Microsecond,
RaftDbWriteMemtableDuration: 50 * time.Microsecond,
CommitLogDuration: 60 * time.Microsecond,
ApplyBatchWaitDuration: 70 * time.Microsecond,
ApplyLogDuration: 80 * time.Microsecond,
ApplyMutexLockDuration: 90 * time.Microsecond,
ApplyWriteLeaderWaitDuration: 100 * time.Microsecond,
ApplyWriteWalDuration: 101 * time.Microsecond,
ApplyWriteMemtableDuration: 102 * time.Microsecond,
},
},
},
CommitPrimary: util.ReqDetailInfo{},
},
WriteKeys: 3,
WriteSize: 66,
PrewriteRegionNum: 5,
TxnRetry: 2,
ResolveLock: util.ResolveLockDetail{
ResolveLockTime: int64(time.Second),
},
}
stats := &RuntimeStatsWithCommit{
Commit: commitDetail,
}
expect := "commit_txn: {prewrite:1s, get_commit_ts:1s, commit:1s, backoff: {time: 1s, prewrite type: [backoff1 backoff2]}, " +
"slowest_prewrite_rpc: {total: 1.000s, region_id: 1000, store: tikv-1:20160, tikv_wall_time: 500ms, " +
"scan_detail: {total_process_keys: 10, total_keys: 100, rocksdb: {delete_skipped_count: 1, key_skipped_count: 1, " +
"block: {cache_hit_count: 1, read_count: 1, read_byte: 100 Bytes, read_time: 20ms}}}, " +
"write_detail: {store_batch_wait: 10µs, propose_send_wait: 20µs, persist_log: {total: 30µs, write_leader_wait: 40µs, " +
"sync_log: 45µs, write_memtable: 50µs}, commit_log: 60µs, apply_batch_wait: 70µs, apply: {total:80µs, mutex_lock: 90µs, " +
"write_leader_wait: 100µs, write_wal: 101µs, write_memtable: 102µs}}}, resolve_lock: 1s, region_num:5, write_keys:3" +
", write_byte:66, txn_retry:2}"
require.Equal(t, expect, stats.String())
lockDetail := &util.LockKeysDetails{
TotalTime: time.Second,
RegionNum: 2,
LockKeys: 10,
BackoffTime: int64(time.Second * 3),
Mu: struct {
sync.Mutex
BackoffTypes []string
SlowestReqTotalTime time.Duration
SlowestRegion uint64
SlowestStoreAddr string
SlowestExecDetails util.TiKVExecDetails
}{
BackoffTypes: []string{
"backoff4",
"backoff5",
"backoff5",
},
SlowestReqTotalTime: time.Second,
SlowestRegion: 1000,
SlowestStoreAddr: "tikv-1:20160",
SlowestExecDetails: util.TiKVExecDetails{
TimeDetail: &util.TimeDetail{
TotalRPCWallTime: 500 * time.Millisecond,
},
ScanDetail: &util.ScanDetail{
ProcessedKeys: 10,
TotalKeys: 100,
RocksdbDeleteSkippedCount: 1,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 1,
RocksdbBlockReadCount: 1,
RocksdbBlockReadByte: 100,
RocksdbBlockReadDuration: 20 * time.Millisecond,
},
WriteDetail: &util.WriteDetail{
StoreBatchWaitDuration: 10 * time.Microsecond,
ProposeSendWaitDuration: 20 * time.Microsecond,
PersistLogDuration: 30 * time.Microsecond,
RaftDbWriteLeaderWaitDuration: 40 * time.Microsecond,
RaftDbSyncLogDuration: 45 * time.Microsecond,
RaftDbWriteMemtableDuration: 50 * time.Microsecond,
CommitLogDuration: 60 * time.Microsecond,
ApplyBatchWaitDuration: 70 * time.Microsecond,
ApplyLogDuration: 80 * time.Microsecond,
ApplyMutexLockDuration: 90 * time.Microsecond,
ApplyWriteLeaderWaitDuration: 100 * time.Microsecond,
ApplyWriteWalDuration: 101 * time.Microsecond,
ApplyWriteMemtableDuration: 102 * time.Microsecond,
},
}},
LockRPCTime: int64(time.Second * 5),
LockRPCCount: 50,
RetryCount: 2,
ResolveLock: util.ResolveLockDetail{
ResolveLockTime: int64(time.Second * 2),
},
}
stats = &RuntimeStatsWithCommit{
LockKeys: lockDetail,
}
expect = "lock_keys: {time:1s, region:2, keys:10, resolve_lock:2s, backoff: {time: 3s, type: [backoff4 backoff5]}, " +
"slowest_rpc: {total: 1.000s, region_id: 1000, store: tikv-1:20160, tikv_wall_time: 500ms, scan_detail: " +
"{total_process_keys: 10, total_keys: 100, rocksdb: {delete_skipped_count: 1, key_skipped_count: 1, block: " +
"{cache_hit_count: 1, read_count: 1, read_byte: 100 Bytes, read_time: 20ms}}}, write_detail: " +
"{store_batch_wait: 10µs, propose_send_wait: 20µs, persist_log: {total: 30µs, write_leader_wait: 40µs, sync_log: 45µs, write_memtable: 50µs}, " +
"commit_log: 60µs, apply_batch_wait: 70µs, apply: {total:80µs, mutex_lock: 90µs, write_leader_wait: 100µs, write_wal: 101µs, write_memtable: 102µs}}}, " +
"lock_rpc:5s, rpc_count:50, retry_count:2}"
require.Equal(t, expect, stats.String())
}
func TestRootRuntimeStats(t *testing.T) {
pid := 1
stmtStats := NewRuntimeStatsColl(nil)
basic1 := stmtStats.GetBasicRuntimeStats(pid)
basic2 := stmtStats.GetBasicRuntimeStats(pid)
basic1.Record(time.Second, 20)
basic2.Record(time.Second*2, 30)
concurrency := &RuntimeStatsWithConcurrencyInfo{}
concurrency.SetConcurrencyInfo(NewConcurrencyInfo("worker", 15))
commitDetail := &util.CommitDetails{
GetCommitTsTime: time.Second,
PrewriteTime: time.Second,
CommitTime: time.Second,
WriteKeys: 3,
WriteSize: 66,
PrewriteRegionNum: 5,
TxnRetry: 2,
}
stmtStats.RegisterStats(pid, concurrency)
stmtStats.RegisterStats(pid, &RuntimeStatsWithCommit{
Commit: commitDetail,
})
stats := stmtStats.GetRootStats(1)
expect := "time:3s, loops:2, worker:15, commit_txn: {prewrite:1s, get_commit_ts:1s, commit:1s, region_num:5, write_keys:3, write_byte:66, txn_retry:2}"
require.Equal(t, expect, stats.String())
}
func TestFormatDurationForExplain(t *testing.T) {
cases := []struct {
t string
s string
}{
{"0s", "0s"},
{"1ns", "1ns"},
{"9ns", "9ns"},
{"10ns", "10ns"},
{"999ns", "999ns"},
{"1µs", "1µs"},
{"1.123µs", "1.12µs"},
{"1.023µs", "1.02µs"},
{"1.003µs", "1µs"},
{"10.456µs", "10.5µs"},
{"10.956µs", "11µs"},
{"999.056µs", "999.1µs"},
{"999.988µs", "1ms"},
{"1.123ms", "1.12ms"},
{"1.023ms", "1.02ms"},
{"1.003ms", "1ms"},
{"10.456ms", "10.5ms"},
{"10.956ms", "11ms"},
{"999.056ms", "999.1ms"},
{"999.988ms", "1s"},
{"1.123s", "1.12s"},
{"1.023s", "1.02s"},
{"1.003s", "1s"},
{"10.456s", "10.5s"},
{"10.956s", "11s"},
{"16m39.056s", "16m39.1s"},
{"16m39.988s", "16m40s"},
{"24h16m39.388662s", "24h16m39.4s"},
{"9.412345ms", "9.41ms"},
{"10.412345ms", "10.4ms"},
{"5.999s", "6s"},
{"100.45µs", "100.5µs"},
}
for _, ca := range cases {
d, err := time.ParseDuration(ca.t)
require.NoError(t, err)
result := FormatDuration(d)
require.Equal(t, ca.s, result)
}
}
func TestCopRuntimeStats2(t *testing.T) {
stats := NewRuntimeStatsColl(nil)
tableScanID := 1
scanDetail := &util.ScanDetail{
TotalKeys: 15,
ProcessedKeys: 10,
ProcessedKeysSize: 10,
RocksdbDeleteSkippedCount: 5,
RocksdbKeySkippedCount: 1,
RocksdbBlockCacheHitCount: 10,
RocksdbBlockReadCount: 20,
RocksdbBlockReadByte: 100,
}
stats.RecordScanDetail(tableScanID, "tikv", scanDetail)
for i := 0; i < 1005; i++ {
stats.RecordOneCopTask(tableScanID, "tikv", "8.8.8.9", mockExecutorExecutionSummary(2, 2, 2))
stats.RecordScanDetail(tableScanID, "tikv", scanDetail)
}
cop := stats.GetOrCreateCopStats(tableScanID, "tikv")
expected := "tikv_task:{proc max:0s, min:0s, avg: 2ns, p80:2ns, p95:2ns, iters:2010, tasks:1005}, " +
"scan_detail: {total_process_keys: 10060, total_process_keys_size: 10060, total_keys: 15090, rocksdb: {delete_skipped_count: 5030, key_skipped_count: 1006, block: {cache_hit_count: 10060, read_count: 20120, read_byte: 98.2 KB}}}"
require.Equal(t, expected, cop.String())
require.Equal(t, expected, cop.String())
}