Files
tidb/br/pkg/streamhelper/client.go
2025-03-03 09:01:57 +00:00

616 lines
18 KiB
Go

// Copyright 2021 PingCAP, Inc. Licensed under Apache-2.0.
package streamhelper
import (
"bytes"
"context"
"encoding/binary"
"encoding/json"
"fmt"
"mime"
"os"
"strconv"
"strings"
"time"
"github.com/gogo/protobuf/proto"
"github.com/pingcap/errors"
backuppb "github.com/pingcap/kvproto/pkg/brpb"
"github.com/pingcap/log"
berrors "github.com/pingcap/tidb/br/pkg/errors"
"github.com/pingcap/tidb/pkg/kv"
"github.com/pingcap/tidb/pkg/util/redact"
clientv3 "go.etcd.io/etcd/client/v3"
"go.uber.org/zap"
)
const (
SeverityError = "ERROR"
SeverityManual = "MANUAL"
)
// RFC3339Time is a wrapper of `time.Time` that marshals to a RFC3339
// JSON string when encoding to / decoding from json.
type RFC3339Time time.Time
// MarshalJSON implements the `json.Marshaler` interface.
func (t RFC3339Time) MarshalJSON() ([]byte, error) {
return []byte(strconv.Quote(t.String())), nil
}
// UnmarshalJSON implements the `json.Unmarshaler` interface.
func (t *RFC3339Time) UnmarshalJSON(data []byte) error {
str, err := strconv.Unquote(string(data))
if err != nil {
return errors.Annotatef(err, "RFC3339Time: the data to unmarshal isn't a json string (%v)", data)
}
tm, err := time.Parse(time.RFC3339, str)
if err != nil {
return errors.Annotatef(err, "RFC3339Time: the data isn't a valid RFC3339 time (%s)", str)
}
*t = RFC3339Time(tm)
return nil
}
func (t RFC3339Time) String() string {
return time.Time(t).Format(time.RFC3339)
}
// NewLocalPauseV2 creates a new `PauseV2` that ready to be attached to a paused task.
func NewLocalPauseV2() *PauseV2 {
pid := os.Getpid()
hostName, err := os.Hostname()
if err != nil {
hostName = fmt.Sprintf("[err: %s]", err)
}
return &PauseV2{
Severity: SeverityManual,
OperatorHostName: hostName,
OperatorPID: pid,
OperationTime: RFC3339Time(time.Now()),
}
}
// PauseV2 is extra information attached to a paused task.
type PauseV2 struct {
// Severity is the severity of this pause.
// `SeverityError`: The task encounters some fatal errors and have to be paused.
// `SeverityRegularOperation`: The task was paused by a normal operation.
Severity string `json:"severity"`
// OperatorHostName is the hostname that pauses the task.
OperatorHostName string `json:"operation_hostname"`
// OperatorPID is the pid of the operator process.
OperatorPID int `json:"operation_pid"`
// OperationTime is the time when the task was paused.
OperationTime RFC3339Time `json:"operation_time"`
// PayloadType is the mime type of the payload.
// For now, only the two types are supported:
// - application/x-protobuf?messagetype=brpb.StreamBackupError
// - text/plain
PayloadType string `json:"payload_type"`
// Payload is the payload attached to the pause.
Payload []byte `json:"payload"`
}
// GetPayload returns the payload attached to the pause.
// The returned value should be one of `string` or `*backuppb.StreamBackupError`.
func (p *PauseV2) GetPayload() (any, error) {
m, param, err := mime.ParseMediaType(p.PayloadType)
if err != nil {
return nil, errors.Annotatef(err, "%s isn't a valid mime type", p.PayloadType)
}
switch m {
case "text/plain":
// Note: consider the charset?
return string(p.Payload), nil
case "application/x-protobuf":
msgType, ok := param["messagetype"]
if !ok {
return nil, errors.Errorf("x-protobuf didn't specified msgType (%s)", p.PayloadType)
}
if msgType != "brpb.StreamBackupError" {
return nil, errors.Errorf("only type brpb.StreamBackupError is supported (%s)", p.PayloadType)
}
var msg backuppb.StreamBackupError
err := proto.Unmarshal(p.Payload, &msg)
if err != nil {
return nil, errors.Annotatef(err, "failed to unmarshal the payload")
}
return &msg, nil
default:
return nil, errors.Errorf("unsupported payload type %s", m)
}
}
// DisplayTable prints human-friendly records to the provided function.
func (p *PauseV2) DisplayTable(display func(string, string)) {
display("pause-time", p.OperationTime.String())
display("pause-operator", p.OperatorHostName)
display("pause-operator-pid", strconv.Itoa(p.OperatorPID))
pl, err := p.GetPayload()
if err != nil {
display("pause-payload[errparse]", err.Error())
} else {
switch val := pl.(type) {
case string:
display("pause-payload", val)
case *backuppb.StreamBackupError:
display("pause-payload[errcode]", val.ErrorCode)
display("pause-payload[errmesg]", val.ErrorMessage)
default:
display("pause-payload", "unknown")
}
}
}
func (p *PauseV2) SetTextMessage(t string) {
p.PayloadType = "text/plain;charset=UTF-8"
p.Payload = []byte(t)
}
func (p *PauseV2) SetBakcupStreamError(berr *backuppb.StreamBackupError) (err error) {
p.PayloadType = "application/x-protobuf;messagetype=brpb.StreamBackupError"
p.Payload, err = proto.Marshal(berr)
return err
}
// Checkpoint is the polymorphic checkpoint type.
// The `ID` and `Version` implies the type of this checkpoint:
// When ID == 0 and Version == 0, it is the task start ts.
// When ID != 0 and Version == 0, it is the store level checkpoint.
// When ID != 0 and Version != 0, it is the region level checkpoint.
type Checkpoint struct {
ID uint64 `json:"id,omitempty"`
Version uint64 `json:"epoch_version,omitempty"`
TS uint64 `json:"ts"`
IsGlobal bool `json:"-"`
}
type CheckpointType int
const (
CheckpointTypeStore CheckpointType = iota
CheckpointTypeRegion
CheckpointTypeTask
CheckpointTypeGlobal
CheckpointTypeInvalid
)
// Type returns the type(provider) of the checkpoint.
func (cp Checkpoint) Type() CheckpointType {
switch {
case cp.IsGlobal:
return CheckpointTypeGlobal
case cp.ID == 0 && cp.Version == 0:
return CheckpointTypeTask
case cp.ID != 0 && cp.Version == 0:
return CheckpointTypeStore
case cp.ID != 0 && cp.Version != 0:
return CheckpointTypeRegion
default:
return CheckpointTypeInvalid
}
}
// MetaDataClient is the client for operations over metadata.
type MetaDataClient struct {
*clientv3.Client
}
func NewMetaDataClient(c *clientv3.Client) *MetaDataClient {
return &MetaDataClient{c}
}
// ParseCheckpoint parses the checkpoint from a key & value pair.
func ParseCheckpoint(task string, key, value []byte) (Checkpoint, error) {
pfx := []byte(CheckPointsOf(task))
if !bytes.HasPrefix(key, pfx) {
return Checkpoint{}, errors.Annotatef(berrors.ErrInvalidArgument, "the prefix is wrong for key: %s", key)
}
key = bytes.TrimPrefix(key, pfx)
segs := bytes.Split(key, []byte("/"))
var checkpoint Checkpoint
switch string(segs[0]) {
case checkpointTypeStore:
if len(segs) != 2 {
return checkpoint, errors.Annotatef(berrors.ErrPiTRMalformedMetadata,
"the store checkpoint seg mismatch; segs = %v", segs)
}
id, err := strconv.ParseUint(string(segs[1]), 10, 64)
if err != nil {
return checkpoint, err
}
checkpoint.ID = id
case checkpointTypeGlobal:
checkpoint.IsGlobal = true
case checkpointTypeRegion:
if len(segs) != 3 {
return checkpoint, errors.Annotatef(berrors.ErrPiTRMalformedMetadata,
"the region checkpoint seg mismatch; segs = %v", segs)
}
id, err := strconv.ParseUint(string(segs[1]), 10, 64)
if err != nil {
return checkpoint, err
}
version, err := strconv.ParseUint(string(segs[2]), 10, 64)
if err != nil {
return checkpoint, err
}
checkpoint.ID = id
checkpoint.Version = version
default:
if len(key) != 8 {
return checkpoint, errors.Annotatef(berrors.ErrPiTRMalformedMetadata,
"the store id isn't 64bits (it is %d bytes, value = %s)",
len(key),
redact.Key(key))
}
id := binary.BigEndian.Uint64(key)
checkpoint.ID = id
}
if len(value) != 8 {
return checkpoint, errors.Annotatef(berrors.ErrPiTRMalformedMetadata,
"the checkpoint value isn't 64bits (it is %d bytes, value = %s)",
len(segs[0]),
redact.Key(segs[0]))
}
checkpoint.TS = binary.BigEndian.Uint64(value)
return checkpoint, nil
}
// PutTask put a task to the metadata storage.
func (c *MetaDataClient) PutTask(ctx context.Context, task TaskInfo) error {
data, err := task.PBInfo.Marshal()
if err != nil {
return errors.Annotatef(err, "failed to marshal task %s", task.PBInfo.Name)
}
ops := make([]clientv3.Op, 0, 2+len(task.Ranges))
ops = append(ops, clientv3.OpPut(TaskOf(task.PBInfo.Name), string(data)))
for _, r := range task.Ranges {
ops = append(ops, clientv3.OpPut(RangeKeyOf(task.PBInfo.Name, r.StartKey), string(r.EndKey)))
log.Debug("range info",
zap.String("task-name", task.PBInfo.Name),
zap.String("start-key", redact.Key(r.StartKey)),
zap.String("end-key", redact.Key(r.EndKey)),
)
}
if task.Pausing {
ops = append(ops, clientv3.OpPut(Pause(task.PBInfo.Name), ""))
}
txn := c.KV.Txn(ctx)
_, err = txn.Then(ops...).Commit()
if err != nil {
return errors.Annotatef(err, "failed to commit the change for task %s", task.PBInfo.Name)
}
return nil
}
// DeleteTask deletes a task, along with its metadata.
func (c *MetaDataClient) DeleteTask(ctx context.Context, taskName string) error {
_, err := c.KV.Txn(ctx).
Then(
clientv3.OpDelete(TaskOf(taskName)),
clientv3.OpDelete(RangesOf(taskName), clientv3.WithPrefix()),
clientv3.OpDelete(CheckPointsOf(taskName), clientv3.WithPrefix()),
clientv3.OpDelete(Pause(taskName)),
clientv3.OpDelete(LastErrorPrefixOf(taskName), clientv3.WithPrefix()),
clientv3.OpDelete(GlobalCheckpointOf(taskName)),
clientv3.OpDelete(StorageCheckpointOf(taskName), clientv3.WithPrefix()),
).
Commit()
if err != nil {
return errors.Annotatef(err, "failed to delete task itself %s", taskName)
}
return nil
}
type pauseTaskCfg struct {
message string
severity string
}
type PauseTaskOption func(*PauseV2)
func PauseWithMessage(m string) PauseTaskOption {
return func(p *PauseV2) {
p.SetTextMessage(m)
}
}
func PauseWithErrorSeverity(p *PauseV2) {
p.Severity = SeverityError
}
func (c *MetaDataClient) PauseTask(ctx context.Context, taskName string, opts ...PauseTaskOption) error {
p := NewLocalPauseV2()
for _, opt := range opts {
opt(p)
}
data, err := json.Marshal(p)
if err != nil {
return errors.Annotatef(err, "failed to marshal pause payload for task %s", taskName)
}
_, err = c.KV.Put(ctx, Pause(taskName), string(data))
if err != nil {
return errors.Annotatef(err, "failed to pause task %s", taskName)
}
return nil
}
func (c *MetaDataClient) ResumeTask(ctx context.Context, taskName string) error {
_, err := c.KV.Delete(ctx, Pause(taskName))
if err != nil {
return errors.Annotatef(err, "failed to resume task %s", taskName)
}
return nil
}
func (c *MetaDataClient) CleanLastErrorOfTask(ctx context.Context, taskName string) error {
_, err := c.KV.Delete(ctx, LastErrorPrefixOf(taskName), clientv3.WithPrefix())
if err != nil {
return errors.Annotatef(err, "failed to clean last error of task %s", taskName)
}
return nil
}
// GetTask get the basic task handle from the metadata storage.
func (c *MetaDataClient) GetTask(ctx context.Context, taskName string) (*Task, error) {
resp, err := c.Get(ctx, TaskOf(taskName))
if err != nil {
return nil, errors.Annotatef(err, "failed to fetch task %s", taskName)
}
if len(resp.Kvs) == 0 {
return nil, errors.Annotatef(berrors.ErrPiTRTaskNotFound, "no such task %s", taskName)
}
var taskInfo backuppb.StreamBackupTaskInfo
err = proto.Unmarshal(resp.Kvs[0].Value, &taskInfo)
if err != nil {
return nil, errors.Annotatef(err, "invalid binary presentation of task info (name = %s)", taskName)
}
task := &Task{
cli: c,
Info: taskInfo,
}
return task, nil
}
func (c *MetaDataClient) GetTaskWithPauseStatus(ctx context.Context, taskName string) (*Task, bool, error) {
resps, err := c.KV.Txn(ctx).
Then(
clientv3.OpGet(TaskOf(taskName)),
clientv3.OpGet(Pause(taskName)),
).Commit()
if err != nil {
return nil, false, errors.Annotatef(err, "failed to fetch task %s", taskName)
}
if len(resps.Responses) == 0 || len(resps.Responses[0].GetResponseRange().Kvs) == 0 {
return nil, false, errors.Annotatef(berrors.ErrPiTRTaskNotFound, "no such task %s", taskName)
}
var taskInfo backuppb.StreamBackupTaskInfo
err = proto.Unmarshal(resps.Responses[0].GetResponseRange().Kvs[0].Value, &taskInfo)
if err != nil {
return nil, false, errors.Annotatef(err, "invalid binary presentation of task info (name = %s)", taskName)
}
paused := false
if len(resps.Responses) > 1 && len(resps.Responses[1].GetResponseRange().Kvs) > 0 {
paused = true
}
return &Task{cli: c, Info: taskInfo}, paused, nil
}
func (c *MetaDataClient) TaskByInfo(t backuppb.StreamBackupTaskInfo) *Task {
return &Task{cli: c, Info: t}
}
func (c *MetaDataClient) GetAllTasksWithRevision(ctx context.Context) ([]Task, int64, error) {
resp, err := c.KV.Get(ctx, PrefixOfTask(), clientv3.WithPrefix())
if err != nil {
return nil, 0, errors.Trace(err)
}
kvs := resp.Kvs
if len(kvs) == 0 {
return nil, resp.Header.GetRevision(), nil
}
tasks := make([]Task, len(kvs))
for idx, kv := range kvs {
err = proto.Unmarshal(kv.Value, &tasks[idx].Info)
if err != nil {
return nil, 0, errors.Trace(err)
}
tasks[idx].cli = c
}
return tasks, resp.Header.GetRevision(), nil
}
// GetAllTasks get all of tasks from metadata storage.
func (c *MetaDataClient) GetAllTasks(ctx context.Context) ([]Task, error) {
tasks, _, err := c.GetAllTasksWithRevision(ctx)
return tasks, err
}
// GetTaskCount get the count of tasks from metadata storage.
func (c *MetaDataClient) GetTaskCount(ctx context.Context) (int, error) {
scanner := scanEtcdPrefix(c.Client, PrefixOfTask())
kvs, err := scanner.AllPages(ctx, 1)
return len(kvs), errors.Trace(err)
}
// Task presents a remote "task" object.
// returned by a query of task.
// Associated to the client created it, hence be able to fetch remote fields like `ranges`.
type Task struct {
cli *MetaDataClient
Info backuppb.StreamBackupTaskInfo
}
func NewTask(client *MetaDataClient, info backuppb.StreamBackupTaskInfo) *Task {
return &Task{
cli: client,
Info: info,
}
}
// Pause is a shorthand for `metaCli.PauseTask`.
func (t *Task) Pause(ctx context.Context, opts ...PauseTaskOption) error {
return t.cli.PauseTask(ctx, t.Info.Name, opts...)
}
// Resume is a shorthand for `metaCli.ResumeTask`
func (t *Task) Resume(ctx context.Context) error {
return t.cli.ResumeTask(ctx, t.Info.Name)
}
func (t *Task) GetPauseV2(ctx context.Context) (*PauseV2, error) {
resp, err := t.cli.KV.Get(ctx, Pause(t.Info.Name))
if err != nil {
return nil, errors.Annotatef(err, "failed to fetch the status of task %s", t.Info.Name)
}
if resp.Count == 0 {
return nil, nil
}
rawPauseV2 := resp.Kvs[0].Value
if len(rawPauseV2) == 0 {
return nil, nil
}
var pauseV2 PauseV2
err = json.Unmarshal(rawPauseV2, &pauseV2)
if err != nil {
return nil, errors.Annotatef(err, "failed to unmarshal pause payload for task %s", t.Info.Name)
}
return &pauseV2, nil
}
func (t *Task) IsPaused(ctx context.Context) (bool, error) {
resp, err := t.cli.KV.Get(ctx, Pause(t.Info.Name), clientv3.WithCountOnly())
if err != nil {
return false, errors.Annotatef(err, "failed to fetch the status of task %s", t.Info.Name)
}
return resp.Count > 0, nil
}
// Ranges tries to fetch the range from the metadata storage.
func (t *Task) Ranges(ctx context.Context) (Ranges, error) {
ranges := make(Ranges, 0, 64)
kvs, err := scanEtcdPrefix(t.cli.Client, RangesOf(t.Info.Name)).AllPages(ctx, 64)
if err != nil {
return nil, errors.Annotatef(err, "failed to fetch ranges of task %s", t.Info.Name)
}
commonPrefix := []byte(RangesOf(t.Info.Name))
for _, kvp := range kvs {
// The prefix must matches.
startKey := kvp.Key[len(commonPrefix):]
ranges = append(ranges, kv.KeyRange{StartKey: startKey, EndKey: kvp.Value})
}
return ranges, nil
}
// NextBackupTSList lists the backup ts of each store.
func (t *Task) NextBackupTSList(ctx context.Context) ([]Checkpoint, error) {
cps := make([]Checkpoint, 0)
prefix := CheckPointsOf(t.Info.Name)
scanner := scanEtcdPrefix(t.cli.Client, prefix)
kvs, err := scanner.AllPages(ctx, 1024)
if err != nil {
return nil, errors.Annotatef(err, "failed to get checkpoints of %s", t.Info.Name)
}
for _, kv := range kvs {
cp, err := ParseCheckpoint(t.Info.Name, kv.Key, kv.Value)
if err != nil {
return cps, err
}
cps = append(cps, cp)
}
return cps, nil
}
func (t *Task) GetStorageCheckpoint(ctx context.Context) (uint64, error) {
prefix := StorageCheckpointOf(t.Info.Name)
scanner := scanEtcdPrefix(t.cli.Client, prefix)
kvs, err := scanner.AllPages(ctx, 1024)
if err != nil {
return 0, errors.Annotatef(err, "failed to get checkpoints of %s", t.Info.Name)
}
var storageCheckpoint = t.Info.StartTs
for _, kv := range kvs {
if len(kv.Value) != 8 {
return 0, errors.Annotatef(berrors.ErrPiTRMalformedMetadata,
"the value isn't 64bits (it is %d bytes, value = %s)",
len(kv.Value),
redact.Key(kv.Value))
}
ts := binary.BigEndian.Uint64(kv.Value)
storageCheckpoint = max(storageCheckpoint, ts)
}
return storageCheckpoint, nil
}
// GetGlobalCheckPointTS gets the global checkpoint timestamp according to log task.
func (t *Task) GetGlobalCheckPointTS(ctx context.Context) (uint64, error) {
checkPointMap, err := t.NextBackupTSList(ctx)
if err != nil {
return 0, errors.Trace(err)
}
initialized := false
checkpoint := t.Info.StartTs
for _, cp := range checkPointMap {
if cp.Type() == CheckpointTypeGlobal {
return cp.TS, nil
}
if cp.Type() == CheckpointTypeStore && (!initialized || cp.TS < checkpoint) {
initialized = true
checkpoint = cp.TS
}
}
ts, err := t.GetStorageCheckpoint(ctx)
if err != nil {
return 0, errors.Trace(err)
}
return max(checkpoint, ts), nil
}
func (t *Task) UploadGlobalCheckpoint(ctx context.Context, ts uint64) error {
_, err := t.cli.KV.Put(ctx, GlobalCheckpointOf(t.Info.Name), string(encodeUint64(ts)))
if err != nil {
return err
}
return nil
}
func (t *Task) LastError(ctx context.Context) (map[uint64]backuppb.StreamBackupError, error) {
storeToError := map[uint64]backuppb.StreamBackupError{}
prefix := LastErrorPrefixOf(t.Info.Name)
result, err := t.cli.Get(ctx, prefix, clientv3.WithPrefix())
if err != nil {
return nil, errors.Annotatef(err, "failed to get the last error for task %s", t.Info.GetName())
}
for _, r := range result.Kvs {
storeIDStr := strings.TrimPrefix(string(r.Key), prefix)
storeID, err := strconv.ParseUint(storeIDStr, 10, 64)
if err != nil {
return nil, errors.Annotatef(err, "failed to parse the store ID string %s", storeIDStr)
}
var lastErr backuppb.StreamBackupError
if err := proto.Unmarshal(r.Value, &lastErr); err != nil {
return nil, errors.Annotatef(err, "failed to parse wire encoding for store %d", storeID)
}
storeToError[storeID] = lastErr
}
return storeToError, nil
}