356 lines
10 KiB
Go
356 lines
10 KiB
Go
// Copyright 2016 PingCAP, Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// Package tikv provides tcp connection to kvserver.
|
|
package tikv
|
|
|
|
import (
|
|
"context"
|
|
"io"
|
|
"math"
|
|
"strconv"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/grpc-ecosystem/go-grpc-middleware/tracing/opentracing"
|
|
"github.com/opentracing/opentracing-go"
|
|
"github.com/pingcap/errors"
|
|
"github.com/pingcap/kvproto/pkg/coprocessor"
|
|
"github.com/pingcap/kvproto/pkg/debugpb"
|
|
"github.com/pingcap/kvproto/pkg/tikvpb"
|
|
"github.com/pingcap/parser/terror"
|
|
"github.com/pingcap/tidb/config"
|
|
"github.com/pingcap/tidb/metrics"
|
|
"github.com/pingcap/tidb/store/tikv/tikvrpc"
|
|
"github.com/pingcap/tidb/util/logutil"
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/credentials"
|
|
"google.golang.org/grpc/keepalive"
|
|
)
|
|
|
|
// MaxRecvMsgSize set max gRPC receive message size received from server. If any message size is larger than
|
|
// current value, an error will be reported from gRPC.
|
|
var MaxRecvMsgSize = math.MaxInt64
|
|
|
|
// Timeout durations.
|
|
var (
|
|
dialTimeout = 5 * time.Second
|
|
readTimeoutShort = 20 * time.Second // For requests that read/write several key-values.
|
|
ReadTimeoutMedium = 60 * time.Second // For requests that may need scan region.
|
|
ReadTimeoutLong = 150 * time.Second // For requests that may need scan region multiple times.
|
|
GCTimeout = 5 * time.Minute
|
|
UnsafeDestroyRangeTimeout = 5 * time.Minute
|
|
)
|
|
|
|
const (
|
|
grpcInitialWindowSize = 1 << 30
|
|
grpcInitialConnWindowSize = 1 << 30
|
|
)
|
|
|
|
// Client is a client that sends RPC.
|
|
// It should not be used after calling Close().
|
|
type Client interface {
|
|
// Close should release all data.
|
|
Close() error
|
|
// SendRequest sends Request.
|
|
SendRequest(ctx context.Context, addr string, req *tikvrpc.Request, timeout time.Duration) (*tikvrpc.Response, error)
|
|
}
|
|
|
|
type connArray struct {
|
|
// The target host.
|
|
target string
|
|
|
|
index uint32
|
|
v []*grpc.ClientConn
|
|
// streamTimeout binds with a background goroutine to process coprocessor streaming timeout.
|
|
streamTimeout chan *tikvrpc.Lease
|
|
// batchConn is not null when batch is enabled.
|
|
*batchConn
|
|
}
|
|
|
|
func newConnArray(maxSize uint, addr string, security config.Security, idleNotify *uint32, done <-chan struct{}) (*connArray, error) {
|
|
a := &connArray{
|
|
index: 0,
|
|
v: make([]*grpc.ClientConn, maxSize),
|
|
streamTimeout: make(chan *tikvrpc.Lease, 1024),
|
|
}
|
|
if err := a.Init(addr, security, idleNotify, done); err != nil {
|
|
return nil, err
|
|
}
|
|
return a, nil
|
|
}
|
|
|
|
func (a *connArray) Init(addr string, security config.Security, idleNotify *uint32, done <-chan struct{}) error {
|
|
a.target = addr
|
|
|
|
opt := grpc.WithInsecure()
|
|
if len(security.ClusterSSLCA) != 0 {
|
|
tlsConfig, err := security.ToTLSConfig()
|
|
if err != nil {
|
|
return errors.Trace(err)
|
|
}
|
|
opt = grpc.WithTransportCredentials(credentials.NewTLS(tlsConfig))
|
|
}
|
|
|
|
cfg := config.GetGlobalConfig()
|
|
var (
|
|
unaryInterceptor grpc.UnaryClientInterceptor
|
|
streamInterceptor grpc.StreamClientInterceptor
|
|
)
|
|
if cfg.OpenTracing.Enable {
|
|
unaryInterceptor = grpc_opentracing.UnaryClientInterceptor()
|
|
streamInterceptor = grpc_opentracing.StreamClientInterceptor()
|
|
}
|
|
|
|
allowBatch := cfg.TiKVClient.MaxBatchSize > 0
|
|
if allowBatch {
|
|
a.batchConn = newBatchConn(uint(len(a.v)), cfg.TiKVClient.MaxBatchSize, idleNotify)
|
|
a.pendingRequests = metrics.TiKVPendingBatchRequests.WithLabelValues(a.target)
|
|
}
|
|
keepAlive := cfg.TiKVClient.GrpcKeepAliveTime
|
|
keepAliveTimeout := cfg.TiKVClient.GrpcKeepAliveTimeout
|
|
for i := range a.v {
|
|
ctx, cancel := context.WithTimeout(context.Background(), dialTimeout)
|
|
conn, err := grpc.DialContext(
|
|
ctx,
|
|
addr,
|
|
opt,
|
|
grpc.WithInitialWindowSize(grpcInitialWindowSize),
|
|
grpc.WithInitialConnWindowSize(grpcInitialConnWindowSize),
|
|
grpc.WithUnaryInterceptor(unaryInterceptor),
|
|
grpc.WithStreamInterceptor(streamInterceptor),
|
|
grpc.WithDefaultCallOptions(grpc.MaxCallRecvMsgSize(MaxRecvMsgSize)),
|
|
grpc.WithBackoffMaxDelay(time.Second*3),
|
|
grpc.WithKeepaliveParams(keepalive.ClientParameters{
|
|
Time: time.Duration(keepAlive) * time.Second,
|
|
Timeout: time.Duration(keepAliveTimeout) * time.Second,
|
|
PermitWithoutStream: true,
|
|
}),
|
|
)
|
|
cancel()
|
|
if err != nil {
|
|
// Cleanup if the initialization fails.
|
|
a.Close()
|
|
return errors.Trace(err)
|
|
}
|
|
a.v[i] = conn
|
|
|
|
if allowBatch {
|
|
// Initialize batch streaming clients.
|
|
tikvClient := tikvpb.NewTikvClient(conn)
|
|
streamClient, err := tikvClient.BatchCommands(context.TODO())
|
|
if err != nil {
|
|
a.Close()
|
|
return errors.Trace(err)
|
|
}
|
|
batchClient := &batchCommandsClient{
|
|
target: a.target,
|
|
conn: conn,
|
|
client: streamClient,
|
|
batched: sync.Map{},
|
|
idAlloc: 0,
|
|
closed: 0,
|
|
}
|
|
a.batchCommandsClients = append(a.batchCommandsClients, batchClient)
|
|
go batchClient.batchRecvLoop(cfg.TiKVClient, &a.tikvTransportLayerLoad)
|
|
}
|
|
}
|
|
go tikvrpc.CheckStreamTimeoutLoop(a.streamTimeout, done)
|
|
if allowBatch {
|
|
go a.batchSendLoop(cfg.TiKVClient)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (a *connArray) Get() *grpc.ClientConn {
|
|
next := atomic.AddUint32(&a.index, 1) % uint32(len(a.v))
|
|
return a.v[next]
|
|
}
|
|
|
|
func (a *connArray) Close() {
|
|
if a.batchConn != nil {
|
|
a.batchConn.Close()
|
|
}
|
|
|
|
for i, c := range a.v {
|
|
if c != nil {
|
|
err := c.Close()
|
|
terror.Log(errors.Trace(err))
|
|
a.v[i] = nil
|
|
}
|
|
}
|
|
}
|
|
|
|
// rpcClient is RPC client struct.
|
|
// TODO: Add flow control between RPC clients in TiDB ond RPC servers in TiKV.
|
|
// Since we use shared client connection to communicate to the same TiKV, it's possible
|
|
// that there are too many concurrent requests which overload the service of TiKV.
|
|
type rpcClient struct {
|
|
sync.RWMutex
|
|
done chan struct{}
|
|
|
|
conns map[string]*connArray
|
|
security config.Security
|
|
|
|
idleNotify uint32
|
|
// Periodically check whether there is any connection that is idle and then close and remove these idle connections.
|
|
// Implement background cleanup.
|
|
isClosed bool
|
|
}
|
|
|
|
func newRPCClient(security config.Security) *rpcClient {
|
|
return &rpcClient{
|
|
done: make(chan struct{}, 1),
|
|
conns: make(map[string]*connArray),
|
|
security: security,
|
|
}
|
|
}
|
|
|
|
// NewTestRPCClient is for some external tests.
|
|
func NewTestRPCClient() Client {
|
|
return newRPCClient(config.Security{})
|
|
}
|
|
|
|
func (c *rpcClient) getConnArray(addr string) (*connArray, error) {
|
|
c.RLock()
|
|
if c.isClosed {
|
|
c.RUnlock()
|
|
return nil, errors.Errorf("rpcClient is closed")
|
|
}
|
|
array, ok := c.conns[addr]
|
|
c.RUnlock()
|
|
if !ok {
|
|
var err error
|
|
array, err = c.createConnArray(addr)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
return array, nil
|
|
}
|
|
|
|
func (c *rpcClient) createConnArray(addr string) (*connArray, error) {
|
|
c.Lock()
|
|
defer c.Unlock()
|
|
array, ok := c.conns[addr]
|
|
if !ok {
|
|
var err error
|
|
connCount := config.GetGlobalConfig().TiKVClient.GrpcConnectionCount
|
|
array, err = newConnArray(connCount, addr, c.security, &c.idleNotify, c.done)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
c.conns[addr] = array
|
|
}
|
|
return array, nil
|
|
}
|
|
|
|
func (c *rpcClient) closeConns() {
|
|
c.Lock()
|
|
if !c.isClosed {
|
|
c.isClosed = true
|
|
// close all connections
|
|
for _, array := range c.conns {
|
|
array.Close()
|
|
}
|
|
}
|
|
c.Unlock()
|
|
}
|
|
|
|
// SendRequest sends a Request to server and receives Response.
|
|
func (c *rpcClient) SendRequest(ctx context.Context, addr string, req *tikvrpc.Request, timeout time.Duration) (*tikvrpc.Response, error) {
|
|
if span := opentracing.SpanFromContext(ctx); span != nil && span.Tracer() != nil {
|
|
span1 := span.Tracer().StartSpan("rpcClient.SendRequest", opentracing.ChildOf(span.Context()))
|
|
defer span1.Finish()
|
|
ctx = opentracing.ContextWithSpan(ctx, span1)
|
|
}
|
|
|
|
start := time.Now()
|
|
reqType := req.Type.String()
|
|
storeID := strconv.FormatUint(req.Context.GetPeer().GetStoreId(), 10)
|
|
defer func() {
|
|
metrics.TiKVSendReqHistogram.WithLabelValues(reqType, storeID).Observe(time.Since(start).Seconds())
|
|
}()
|
|
|
|
if atomic.CompareAndSwapUint32(&c.idleNotify, 1, 0) {
|
|
c.recycleIdleConnArray()
|
|
}
|
|
|
|
connArray, err := c.getConnArray(addr)
|
|
if err != nil {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
|
|
if config.GetGlobalConfig().TiKVClient.MaxBatchSize > 0 {
|
|
if batchReq := req.ToBatchCommandsRequest(); batchReq != nil {
|
|
return sendBatchRequest(ctx, addr, connArray.batchConn, batchReq, timeout)
|
|
}
|
|
}
|
|
|
|
if req.IsDebugReq() {
|
|
client := debugpb.NewDebugClient(connArray.Get())
|
|
ctx1, cancel := context.WithTimeout(ctx, timeout)
|
|
defer cancel()
|
|
return tikvrpc.CallDebugRPC(ctx1, client, req)
|
|
}
|
|
|
|
client := tikvpb.NewTikvClient(connArray.Get())
|
|
|
|
if req.Type != tikvrpc.CmdCopStream {
|
|
ctx1, cancel := context.WithTimeout(ctx, timeout)
|
|
defer cancel()
|
|
return tikvrpc.CallRPC(ctx1, client, req)
|
|
}
|
|
|
|
// Coprocessor streaming request.
|
|
// Use context to support timeout for grpc streaming client.
|
|
ctx1, cancel := context.WithCancel(ctx)
|
|
// Should NOT call defer cancel() here because it will cancel further stream.Recv()
|
|
// We put it in copStream.Lease.Cancel call this cancel at copStream.Close
|
|
// TODO: add unit test for SendRequest.
|
|
resp, err := tikvrpc.CallRPC(ctx1, client, req)
|
|
if err != nil {
|
|
cancel()
|
|
return nil, errors.Trace(err)
|
|
}
|
|
|
|
// Put the lease object to the timeout channel, so it would be checked periodically.
|
|
copStream := resp.Resp.(*tikvrpc.CopStreamResponse)
|
|
copStream.Timeout = timeout
|
|
copStream.Lease.Cancel = cancel
|
|
connArray.streamTimeout <- &copStream.Lease
|
|
|
|
// Read the first streaming response to get CopStreamResponse.
|
|
// This can make error handling much easier, because SendReq() retry on
|
|
// region error automatically.
|
|
var first *coprocessor.Response
|
|
first, err = copStream.Recv()
|
|
if err != nil {
|
|
if errors.Cause(err) != io.EOF {
|
|
return nil, errors.Trace(err)
|
|
}
|
|
logutil.BgLogger().Debug("copstream returns nothing for the request.")
|
|
}
|
|
copStream.Response = first
|
|
return resp, nil
|
|
}
|
|
|
|
func (c *rpcClient) Close() error {
|
|
// TODO: add a unit test for SendRequest After Closed
|
|
close(c.done)
|
|
c.closeConns()
|
|
return nil
|
|
}
|