You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1750 lines
36 KiB
1750 lines
36 KiB
package redis |
|
|
|
import ( |
|
"context" |
|
"crypto/tls" |
|
"fmt" |
|
"math" |
|
"net" |
|
"runtime" |
|
"sort" |
|
"sync" |
|
"sync/atomic" |
|
"time" |
|
|
|
"github.com/go-redis/redis/v8/internal" |
|
"github.com/go-redis/redis/v8/internal/hashtag" |
|
"github.com/go-redis/redis/v8/internal/pool" |
|
"github.com/go-redis/redis/v8/internal/proto" |
|
"github.com/go-redis/redis/v8/internal/rand" |
|
) |
|
|
|
var errClusterNoNodes = fmt.Errorf("redis: cluster has no nodes") |
|
|
|
// ClusterOptions are used to configure a cluster client and should be |
|
// passed to NewClusterClient. |
|
type ClusterOptions struct { |
|
// A seed list of host:port addresses of cluster nodes. |
|
Addrs []string |
|
|
|
// NewClient creates a cluster node client with provided name and options. |
|
NewClient func(opt *Options) *Client |
|
|
|
// The maximum number of retries before giving up. Command is retried |
|
// on network errors and MOVED/ASK redirects. |
|
// Default is 3 retries. |
|
MaxRedirects int |
|
|
|
// Enables read-only commands on slave nodes. |
|
ReadOnly bool |
|
// Allows routing read-only commands to the closest master or slave node. |
|
// It automatically enables ReadOnly. |
|
RouteByLatency bool |
|
// Allows routing read-only commands to the random master or slave node. |
|
// It automatically enables ReadOnly. |
|
RouteRandomly bool |
|
|
|
// Optional function that returns cluster slots information. |
|
// It is useful to manually create cluster of standalone Redis servers |
|
// and load-balance read/write operations between master and slaves. |
|
// It can use service like ZooKeeper to maintain configuration information |
|
// and Cluster.ReloadState to manually trigger state reloading. |
|
ClusterSlots func(context.Context) ([]ClusterSlot, error) |
|
|
|
// Following options are copied from Options struct. |
|
|
|
Dialer func(ctx context.Context, network, addr string) (net.Conn, error) |
|
|
|
OnConnect func(ctx context.Context, cn *Conn) error |
|
|
|
Username string |
|
Password string |
|
|
|
MaxRetries int |
|
MinRetryBackoff time.Duration |
|
MaxRetryBackoff time.Duration |
|
|
|
DialTimeout time.Duration |
|
ReadTimeout time.Duration |
|
WriteTimeout time.Duration |
|
|
|
// PoolFIFO uses FIFO mode for each node connection pool GET/PUT (default LIFO). |
|
PoolFIFO bool |
|
|
|
// PoolSize applies per cluster node and not for the whole cluster. |
|
PoolSize int |
|
MinIdleConns int |
|
MaxConnAge time.Duration |
|
PoolTimeout time.Duration |
|
IdleTimeout time.Duration |
|
IdleCheckFrequency time.Duration |
|
|
|
TLSConfig *tls.Config |
|
} |
|
|
|
func (opt *ClusterOptions) init() { |
|
if opt.MaxRedirects == -1 { |
|
opt.MaxRedirects = 0 |
|
} else if opt.MaxRedirects == 0 { |
|
opt.MaxRedirects = 3 |
|
} |
|
|
|
if opt.RouteByLatency || opt.RouteRandomly { |
|
opt.ReadOnly = true |
|
} |
|
|
|
if opt.PoolSize == 0 { |
|
opt.PoolSize = 5 * runtime.GOMAXPROCS(0) |
|
} |
|
|
|
switch opt.ReadTimeout { |
|
case -1: |
|
opt.ReadTimeout = 0 |
|
case 0: |
|
opt.ReadTimeout = 3 * time.Second |
|
} |
|
switch opt.WriteTimeout { |
|
case -1: |
|
opt.WriteTimeout = 0 |
|
case 0: |
|
opt.WriteTimeout = opt.ReadTimeout |
|
} |
|
|
|
if opt.MaxRetries == 0 { |
|
opt.MaxRetries = -1 |
|
} |
|
switch opt.MinRetryBackoff { |
|
case -1: |
|
opt.MinRetryBackoff = 0 |
|
case 0: |
|
opt.MinRetryBackoff = 8 * time.Millisecond |
|
} |
|
switch opt.MaxRetryBackoff { |
|
case -1: |
|
opt.MaxRetryBackoff = 0 |
|
case 0: |
|
opt.MaxRetryBackoff = 512 * time.Millisecond |
|
} |
|
|
|
if opt.NewClient == nil { |
|
opt.NewClient = NewClient |
|
} |
|
} |
|
|
|
func (opt *ClusterOptions) clientOptions() *Options { |
|
const disableIdleCheck = -1 |
|
|
|
return &Options{ |
|
Dialer: opt.Dialer, |
|
OnConnect: opt.OnConnect, |
|
|
|
Username: opt.Username, |
|
Password: opt.Password, |
|
|
|
MaxRetries: opt.MaxRetries, |
|
MinRetryBackoff: opt.MinRetryBackoff, |
|
MaxRetryBackoff: opt.MaxRetryBackoff, |
|
|
|
DialTimeout: opt.DialTimeout, |
|
ReadTimeout: opt.ReadTimeout, |
|
WriteTimeout: opt.WriteTimeout, |
|
|
|
PoolFIFO: opt.PoolFIFO, |
|
PoolSize: opt.PoolSize, |
|
MinIdleConns: opt.MinIdleConns, |
|
MaxConnAge: opt.MaxConnAge, |
|
PoolTimeout: opt.PoolTimeout, |
|
IdleTimeout: opt.IdleTimeout, |
|
IdleCheckFrequency: disableIdleCheck, |
|
|
|
TLSConfig: opt.TLSConfig, |
|
// If ClusterSlots is populated, then we probably have an artificial |
|
// cluster whose nodes are not in clustering mode (otherwise there isn't |
|
// much use for ClusterSlots config). This means we cannot execute the |
|
// READONLY command against that node -- setting readOnly to false in such |
|
// situations in the options below will prevent that from happening. |
|
readOnly: opt.ReadOnly && opt.ClusterSlots == nil, |
|
} |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type clusterNode struct { |
|
Client *Client |
|
|
|
latency uint32 // atomic |
|
generation uint32 // atomic |
|
failing uint32 // atomic |
|
} |
|
|
|
func newClusterNode(clOpt *ClusterOptions, addr string) *clusterNode { |
|
opt := clOpt.clientOptions() |
|
opt.Addr = addr |
|
node := clusterNode{ |
|
Client: clOpt.NewClient(opt), |
|
} |
|
|
|
node.latency = math.MaxUint32 |
|
if clOpt.RouteByLatency { |
|
go node.updateLatency() |
|
} |
|
|
|
return &node |
|
} |
|
|
|
func (n *clusterNode) String() string { |
|
return n.Client.String() |
|
} |
|
|
|
func (n *clusterNode) Close() error { |
|
return n.Client.Close() |
|
} |
|
|
|
func (n *clusterNode) updateLatency() { |
|
const numProbe = 10 |
|
var dur uint64 |
|
|
|
for i := 0; i < numProbe; i++ { |
|
time.Sleep(time.Duration(10+rand.Intn(10)) * time.Millisecond) |
|
|
|
start := time.Now() |
|
n.Client.Ping(context.TODO()) |
|
dur += uint64(time.Since(start) / time.Microsecond) |
|
} |
|
|
|
latency := float64(dur) / float64(numProbe) |
|
atomic.StoreUint32(&n.latency, uint32(latency+0.5)) |
|
} |
|
|
|
func (n *clusterNode) Latency() time.Duration { |
|
latency := atomic.LoadUint32(&n.latency) |
|
return time.Duration(latency) * time.Microsecond |
|
} |
|
|
|
func (n *clusterNode) MarkAsFailing() { |
|
atomic.StoreUint32(&n.failing, uint32(time.Now().Unix())) |
|
} |
|
|
|
func (n *clusterNode) Failing() bool { |
|
const timeout = 15 // 15 seconds |
|
|
|
failing := atomic.LoadUint32(&n.failing) |
|
if failing == 0 { |
|
return false |
|
} |
|
if time.Now().Unix()-int64(failing) < timeout { |
|
return true |
|
} |
|
atomic.StoreUint32(&n.failing, 0) |
|
return false |
|
} |
|
|
|
func (n *clusterNode) Generation() uint32 { |
|
return atomic.LoadUint32(&n.generation) |
|
} |
|
|
|
func (n *clusterNode) SetGeneration(gen uint32) { |
|
for { |
|
v := atomic.LoadUint32(&n.generation) |
|
if gen < v || atomic.CompareAndSwapUint32(&n.generation, v, gen) { |
|
break |
|
} |
|
} |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type clusterNodes struct { |
|
opt *ClusterOptions |
|
|
|
mu sync.RWMutex |
|
addrs []string |
|
nodes map[string]*clusterNode |
|
activeAddrs []string |
|
closed bool |
|
|
|
_generation uint32 // atomic |
|
} |
|
|
|
func newClusterNodes(opt *ClusterOptions) *clusterNodes { |
|
return &clusterNodes{ |
|
opt: opt, |
|
|
|
addrs: opt.Addrs, |
|
nodes: make(map[string]*clusterNode), |
|
} |
|
} |
|
|
|
func (c *clusterNodes) Close() error { |
|
c.mu.Lock() |
|
defer c.mu.Unlock() |
|
|
|
if c.closed { |
|
return nil |
|
} |
|
c.closed = true |
|
|
|
var firstErr error |
|
for _, node := range c.nodes { |
|
if err := node.Client.Close(); err != nil && firstErr == nil { |
|
firstErr = err |
|
} |
|
} |
|
|
|
c.nodes = nil |
|
c.activeAddrs = nil |
|
|
|
return firstErr |
|
} |
|
|
|
func (c *clusterNodes) Addrs() ([]string, error) { |
|
var addrs []string |
|
|
|
c.mu.RLock() |
|
closed := c.closed //nolint:ifshort |
|
if !closed { |
|
if len(c.activeAddrs) > 0 { |
|
addrs = c.activeAddrs |
|
} else { |
|
addrs = c.addrs |
|
} |
|
} |
|
c.mu.RUnlock() |
|
|
|
if closed { |
|
return nil, pool.ErrClosed |
|
} |
|
if len(addrs) == 0 { |
|
return nil, errClusterNoNodes |
|
} |
|
return addrs, nil |
|
} |
|
|
|
func (c *clusterNodes) NextGeneration() uint32 { |
|
return atomic.AddUint32(&c._generation, 1) |
|
} |
|
|
|
// GC removes unused nodes. |
|
func (c *clusterNodes) GC(generation uint32) { |
|
//nolint:prealloc |
|
var collected []*clusterNode |
|
|
|
c.mu.Lock() |
|
|
|
c.activeAddrs = c.activeAddrs[:0] |
|
for addr, node := range c.nodes { |
|
if node.Generation() >= generation { |
|
c.activeAddrs = append(c.activeAddrs, addr) |
|
if c.opt.RouteByLatency { |
|
go node.updateLatency() |
|
} |
|
continue |
|
} |
|
|
|
delete(c.nodes, addr) |
|
collected = append(collected, node) |
|
} |
|
|
|
c.mu.Unlock() |
|
|
|
for _, node := range collected { |
|
_ = node.Client.Close() |
|
} |
|
} |
|
|
|
func (c *clusterNodes) GetOrCreate(addr string) (*clusterNode, error) { |
|
node, err := c.get(addr) |
|
if err != nil { |
|
return nil, err |
|
} |
|
if node != nil { |
|
return node, nil |
|
} |
|
|
|
c.mu.Lock() |
|
defer c.mu.Unlock() |
|
|
|
if c.closed { |
|
return nil, pool.ErrClosed |
|
} |
|
|
|
node, ok := c.nodes[addr] |
|
if ok { |
|
return node, nil |
|
} |
|
|
|
node = newClusterNode(c.opt, addr) |
|
|
|
c.addrs = appendIfNotExists(c.addrs, addr) |
|
c.nodes[addr] = node |
|
|
|
return node, nil |
|
} |
|
|
|
func (c *clusterNodes) get(addr string) (*clusterNode, error) { |
|
var node *clusterNode |
|
var err error |
|
c.mu.RLock() |
|
if c.closed { |
|
err = pool.ErrClosed |
|
} else { |
|
node = c.nodes[addr] |
|
} |
|
c.mu.RUnlock() |
|
return node, err |
|
} |
|
|
|
func (c *clusterNodes) All() ([]*clusterNode, error) { |
|
c.mu.RLock() |
|
defer c.mu.RUnlock() |
|
|
|
if c.closed { |
|
return nil, pool.ErrClosed |
|
} |
|
|
|
cp := make([]*clusterNode, 0, len(c.nodes)) |
|
for _, node := range c.nodes { |
|
cp = append(cp, node) |
|
} |
|
return cp, nil |
|
} |
|
|
|
func (c *clusterNodes) Random() (*clusterNode, error) { |
|
addrs, err := c.Addrs() |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
n := rand.Intn(len(addrs)) |
|
return c.GetOrCreate(addrs[n]) |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type clusterSlot struct { |
|
start, end int |
|
nodes []*clusterNode |
|
} |
|
|
|
type clusterSlotSlice []*clusterSlot |
|
|
|
func (p clusterSlotSlice) Len() int { |
|
return len(p) |
|
} |
|
|
|
func (p clusterSlotSlice) Less(i, j int) bool { |
|
return p[i].start < p[j].start |
|
} |
|
|
|
func (p clusterSlotSlice) Swap(i, j int) { |
|
p[i], p[j] = p[j], p[i] |
|
} |
|
|
|
type clusterState struct { |
|
nodes *clusterNodes |
|
Masters []*clusterNode |
|
Slaves []*clusterNode |
|
|
|
slots []*clusterSlot |
|
|
|
generation uint32 |
|
createdAt time.Time |
|
} |
|
|
|
func newClusterState( |
|
nodes *clusterNodes, slots []ClusterSlot, origin string, |
|
) (*clusterState, error) { |
|
c := clusterState{ |
|
nodes: nodes, |
|
|
|
slots: make([]*clusterSlot, 0, len(slots)), |
|
|
|
generation: nodes.NextGeneration(), |
|
createdAt: time.Now(), |
|
} |
|
|
|
originHost, _, _ := net.SplitHostPort(origin) |
|
isLoopbackOrigin := isLoopback(originHost) |
|
|
|
for _, slot := range slots { |
|
var nodes []*clusterNode |
|
for i, slotNode := range slot.Nodes { |
|
addr := slotNode.Addr |
|
if !isLoopbackOrigin { |
|
addr = replaceLoopbackHost(addr, originHost) |
|
} |
|
|
|
node, err := c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
node.SetGeneration(c.generation) |
|
nodes = append(nodes, node) |
|
|
|
if i == 0 { |
|
c.Masters = appendUniqueNode(c.Masters, node) |
|
} else { |
|
c.Slaves = appendUniqueNode(c.Slaves, node) |
|
} |
|
} |
|
|
|
c.slots = append(c.slots, &clusterSlot{ |
|
start: slot.Start, |
|
end: slot.End, |
|
nodes: nodes, |
|
}) |
|
} |
|
|
|
sort.Sort(clusterSlotSlice(c.slots)) |
|
|
|
time.AfterFunc(time.Minute, func() { |
|
nodes.GC(c.generation) |
|
}) |
|
|
|
return &c, nil |
|
} |
|
|
|
func replaceLoopbackHost(nodeAddr, originHost string) string { |
|
nodeHost, nodePort, err := net.SplitHostPort(nodeAddr) |
|
if err != nil { |
|
return nodeAddr |
|
} |
|
|
|
nodeIP := net.ParseIP(nodeHost) |
|
if nodeIP == nil { |
|
return nodeAddr |
|
} |
|
|
|
if !nodeIP.IsLoopback() { |
|
return nodeAddr |
|
} |
|
|
|
// Use origin host which is not loopback and node port. |
|
return net.JoinHostPort(originHost, nodePort) |
|
} |
|
|
|
func isLoopback(host string) bool { |
|
ip := net.ParseIP(host) |
|
if ip == nil { |
|
return true |
|
} |
|
return ip.IsLoopback() |
|
} |
|
|
|
func (c *clusterState) slotMasterNode(slot int) (*clusterNode, error) { |
|
nodes := c.slotNodes(slot) |
|
if len(nodes) > 0 { |
|
return nodes[0], nil |
|
} |
|
return c.nodes.Random() |
|
} |
|
|
|
func (c *clusterState) slotSlaveNode(slot int) (*clusterNode, error) { |
|
nodes := c.slotNodes(slot) |
|
switch len(nodes) { |
|
case 0: |
|
return c.nodes.Random() |
|
case 1: |
|
return nodes[0], nil |
|
case 2: |
|
if slave := nodes[1]; !slave.Failing() { |
|
return slave, nil |
|
} |
|
return nodes[0], nil |
|
default: |
|
var slave *clusterNode |
|
for i := 0; i < 10; i++ { |
|
n := rand.Intn(len(nodes)-1) + 1 |
|
slave = nodes[n] |
|
if !slave.Failing() { |
|
return slave, nil |
|
} |
|
} |
|
|
|
// All slaves are loading - use master. |
|
return nodes[0], nil |
|
} |
|
} |
|
|
|
func (c *clusterState) slotClosestNode(slot int) (*clusterNode, error) { |
|
nodes := c.slotNodes(slot) |
|
if len(nodes) == 0 { |
|
return c.nodes.Random() |
|
} |
|
|
|
var node *clusterNode |
|
for _, n := range nodes { |
|
if n.Failing() { |
|
continue |
|
} |
|
if node == nil || n.Latency() < node.Latency() { |
|
node = n |
|
} |
|
} |
|
if node != nil { |
|
return node, nil |
|
} |
|
|
|
// If all nodes are failing - return random node |
|
return c.nodes.Random() |
|
} |
|
|
|
func (c *clusterState) slotRandomNode(slot int) (*clusterNode, error) { |
|
nodes := c.slotNodes(slot) |
|
if len(nodes) == 0 { |
|
return c.nodes.Random() |
|
} |
|
if len(nodes) == 1 { |
|
return nodes[0], nil |
|
} |
|
randomNodes := rand.Perm(len(nodes)) |
|
for _, idx := range randomNodes { |
|
if node := nodes[idx]; !node.Failing() { |
|
return node, nil |
|
} |
|
} |
|
return nodes[randomNodes[0]], nil |
|
} |
|
|
|
func (c *clusterState) slotNodes(slot int) []*clusterNode { |
|
i := sort.Search(len(c.slots), func(i int) bool { |
|
return c.slots[i].end >= slot |
|
}) |
|
if i >= len(c.slots) { |
|
return nil |
|
} |
|
x := c.slots[i] |
|
if slot >= x.start && slot <= x.end { |
|
return x.nodes |
|
} |
|
return nil |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type clusterStateHolder struct { |
|
load func(ctx context.Context) (*clusterState, error) |
|
|
|
state atomic.Value |
|
reloading uint32 // atomic |
|
} |
|
|
|
func newClusterStateHolder(fn func(ctx context.Context) (*clusterState, error)) *clusterStateHolder { |
|
return &clusterStateHolder{ |
|
load: fn, |
|
} |
|
} |
|
|
|
func (c *clusterStateHolder) Reload(ctx context.Context) (*clusterState, error) { |
|
state, err := c.load(ctx) |
|
if err != nil { |
|
return nil, err |
|
} |
|
c.state.Store(state) |
|
return state, nil |
|
} |
|
|
|
func (c *clusterStateHolder) LazyReload() { |
|
if !atomic.CompareAndSwapUint32(&c.reloading, 0, 1) { |
|
return |
|
} |
|
go func() { |
|
defer atomic.StoreUint32(&c.reloading, 0) |
|
|
|
_, err := c.Reload(context.Background()) |
|
if err != nil { |
|
return |
|
} |
|
time.Sleep(200 * time.Millisecond) |
|
}() |
|
} |
|
|
|
func (c *clusterStateHolder) Get(ctx context.Context) (*clusterState, error) { |
|
v := c.state.Load() |
|
if v == nil { |
|
return c.Reload(ctx) |
|
} |
|
|
|
state := v.(*clusterState) |
|
if time.Since(state.createdAt) > 10*time.Second { |
|
c.LazyReload() |
|
} |
|
return state, nil |
|
} |
|
|
|
func (c *clusterStateHolder) ReloadOrGet(ctx context.Context) (*clusterState, error) { |
|
state, err := c.Reload(ctx) |
|
if err == nil { |
|
return state, nil |
|
} |
|
return c.Get(ctx) |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type clusterClient struct { |
|
opt *ClusterOptions |
|
nodes *clusterNodes |
|
state *clusterStateHolder //nolint:structcheck |
|
cmdsInfoCache *cmdsInfoCache //nolint:structcheck |
|
} |
|
|
|
// ClusterClient is a Redis Cluster client representing a pool of zero |
|
// or more underlying connections. It's safe for concurrent use by |
|
// multiple goroutines. |
|
type ClusterClient struct { |
|
*clusterClient |
|
cmdable |
|
hooks |
|
ctx context.Context |
|
} |
|
|
|
// NewClusterClient returns a Redis Cluster client as described in |
|
// http://redis.io/topics/cluster-spec. |
|
func NewClusterClient(opt *ClusterOptions) *ClusterClient { |
|
opt.init() |
|
|
|
c := &ClusterClient{ |
|
clusterClient: &clusterClient{ |
|
opt: opt, |
|
nodes: newClusterNodes(opt), |
|
}, |
|
ctx: context.Background(), |
|
} |
|
c.state = newClusterStateHolder(c.loadState) |
|
c.cmdsInfoCache = newCmdsInfoCache(c.cmdsInfo) |
|
c.cmdable = c.Process |
|
|
|
if opt.IdleCheckFrequency > 0 { |
|
go c.reaper(opt.IdleCheckFrequency) |
|
} |
|
|
|
return c |
|
} |
|
|
|
func (c *ClusterClient) Context() context.Context { |
|
return c.ctx |
|
} |
|
|
|
func (c *ClusterClient) WithContext(ctx context.Context) *ClusterClient { |
|
if ctx == nil { |
|
panic("nil context") |
|
} |
|
clone := *c |
|
clone.cmdable = clone.Process |
|
clone.hooks.lock() |
|
clone.ctx = ctx |
|
return &clone |
|
} |
|
|
|
// Options returns read-only Options that were used to create the client. |
|
func (c *ClusterClient) Options() *ClusterOptions { |
|
return c.opt |
|
} |
|
|
|
// ReloadState reloads cluster state. If available it calls ClusterSlots func |
|
// to get cluster slots information. |
|
func (c *ClusterClient) ReloadState(ctx context.Context) { |
|
c.state.LazyReload() |
|
} |
|
|
|
// Close closes the cluster client, releasing any open resources. |
|
// |
|
// It is rare to Close a ClusterClient, as the ClusterClient is meant |
|
// to be long-lived and shared between many goroutines. |
|
func (c *ClusterClient) Close() error { |
|
return c.nodes.Close() |
|
} |
|
|
|
// Do creates a Cmd from the args and processes the cmd. |
|
func (c *ClusterClient) Do(ctx context.Context, args ...interface{}) *Cmd { |
|
cmd := NewCmd(ctx, args...) |
|
_ = c.Process(ctx, cmd) |
|
return cmd |
|
} |
|
|
|
func (c *ClusterClient) Process(ctx context.Context, cmd Cmder) error { |
|
return c.hooks.process(ctx, cmd, c.process) |
|
} |
|
|
|
func (c *ClusterClient) process(ctx context.Context, cmd Cmder) error { |
|
cmdInfo := c.cmdInfo(cmd.Name()) |
|
slot := c.cmdSlot(cmd) |
|
|
|
var node *clusterNode |
|
var ask bool |
|
var lastErr error |
|
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ { |
|
if attempt > 0 { |
|
if err := internal.Sleep(ctx, c.retryBackoff(attempt)); err != nil { |
|
return err |
|
} |
|
} |
|
|
|
if node == nil { |
|
var err error |
|
node, err = c.cmdNode(ctx, cmdInfo, slot) |
|
if err != nil { |
|
return err |
|
} |
|
} |
|
|
|
if ask { |
|
pipe := node.Client.Pipeline() |
|
_ = pipe.Process(ctx, NewCmd(ctx, "asking")) |
|
_ = pipe.Process(ctx, cmd) |
|
_, lastErr = pipe.Exec(ctx) |
|
_ = pipe.Close() |
|
ask = false |
|
} else { |
|
lastErr = node.Client.Process(ctx, cmd) |
|
} |
|
|
|
// If there is no error - we are done. |
|
if lastErr == nil { |
|
return nil |
|
} |
|
if isReadOnly := isReadOnlyError(lastErr); isReadOnly || lastErr == pool.ErrClosed { |
|
if isReadOnly { |
|
c.state.LazyReload() |
|
} |
|
node = nil |
|
continue |
|
} |
|
|
|
// If slave is loading - pick another node. |
|
if c.opt.ReadOnly && isLoadingError(lastErr) { |
|
node.MarkAsFailing() |
|
node = nil |
|
continue |
|
} |
|
|
|
var moved bool |
|
var addr string |
|
moved, ask, addr = isMovedError(lastErr) |
|
if moved || ask { |
|
c.state.LazyReload() |
|
|
|
var err error |
|
node, err = c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
return err |
|
} |
|
continue |
|
} |
|
|
|
if shouldRetry(lastErr, cmd.readTimeout() == nil) { |
|
// First retry the same node. |
|
if attempt == 0 { |
|
continue |
|
} |
|
|
|
// Second try another node. |
|
node.MarkAsFailing() |
|
node = nil |
|
continue |
|
} |
|
|
|
return lastErr |
|
} |
|
return lastErr |
|
} |
|
|
|
// ForEachMaster concurrently calls the fn on each master node in the cluster. |
|
// It returns the first error if any. |
|
func (c *ClusterClient) ForEachMaster( |
|
ctx context.Context, |
|
fn func(ctx context.Context, client *Client) error, |
|
) error { |
|
state, err := c.state.ReloadOrGet(ctx) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
var wg sync.WaitGroup |
|
errCh := make(chan error, 1) |
|
|
|
for _, master := range state.Masters { |
|
wg.Add(1) |
|
go func(node *clusterNode) { |
|
defer wg.Done() |
|
err := fn(ctx, node.Client) |
|
if err != nil { |
|
select { |
|
case errCh <- err: |
|
default: |
|
} |
|
} |
|
}(master) |
|
} |
|
|
|
wg.Wait() |
|
|
|
select { |
|
case err := <-errCh: |
|
return err |
|
default: |
|
return nil |
|
} |
|
} |
|
|
|
// ForEachSlave concurrently calls the fn on each slave node in the cluster. |
|
// It returns the first error if any. |
|
func (c *ClusterClient) ForEachSlave( |
|
ctx context.Context, |
|
fn func(ctx context.Context, client *Client) error, |
|
) error { |
|
state, err := c.state.ReloadOrGet(ctx) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
var wg sync.WaitGroup |
|
errCh := make(chan error, 1) |
|
|
|
for _, slave := range state.Slaves { |
|
wg.Add(1) |
|
go func(node *clusterNode) { |
|
defer wg.Done() |
|
err := fn(ctx, node.Client) |
|
if err != nil { |
|
select { |
|
case errCh <- err: |
|
default: |
|
} |
|
} |
|
}(slave) |
|
} |
|
|
|
wg.Wait() |
|
|
|
select { |
|
case err := <-errCh: |
|
return err |
|
default: |
|
return nil |
|
} |
|
} |
|
|
|
// ForEachShard concurrently calls the fn on each known node in the cluster. |
|
// It returns the first error if any. |
|
func (c *ClusterClient) ForEachShard( |
|
ctx context.Context, |
|
fn func(ctx context.Context, client *Client) error, |
|
) error { |
|
state, err := c.state.ReloadOrGet(ctx) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
var wg sync.WaitGroup |
|
errCh := make(chan error, 1) |
|
|
|
worker := func(node *clusterNode) { |
|
defer wg.Done() |
|
err := fn(ctx, node.Client) |
|
if err != nil { |
|
select { |
|
case errCh <- err: |
|
default: |
|
} |
|
} |
|
} |
|
|
|
for _, node := range state.Masters { |
|
wg.Add(1) |
|
go worker(node) |
|
} |
|
for _, node := range state.Slaves { |
|
wg.Add(1) |
|
go worker(node) |
|
} |
|
|
|
wg.Wait() |
|
|
|
select { |
|
case err := <-errCh: |
|
return err |
|
default: |
|
return nil |
|
} |
|
} |
|
|
|
// PoolStats returns accumulated connection pool stats. |
|
func (c *ClusterClient) PoolStats() *PoolStats { |
|
var acc PoolStats |
|
|
|
state, _ := c.state.Get(context.TODO()) |
|
if state == nil { |
|
return &acc |
|
} |
|
|
|
for _, node := range state.Masters { |
|
s := node.Client.connPool.Stats() |
|
acc.Hits += s.Hits |
|
acc.Misses += s.Misses |
|
acc.Timeouts += s.Timeouts |
|
|
|
acc.TotalConns += s.TotalConns |
|
acc.IdleConns += s.IdleConns |
|
acc.StaleConns += s.StaleConns |
|
} |
|
|
|
for _, node := range state.Slaves { |
|
s := node.Client.connPool.Stats() |
|
acc.Hits += s.Hits |
|
acc.Misses += s.Misses |
|
acc.Timeouts += s.Timeouts |
|
|
|
acc.TotalConns += s.TotalConns |
|
acc.IdleConns += s.IdleConns |
|
acc.StaleConns += s.StaleConns |
|
} |
|
|
|
return &acc |
|
} |
|
|
|
func (c *ClusterClient) loadState(ctx context.Context) (*clusterState, error) { |
|
if c.opt.ClusterSlots != nil { |
|
slots, err := c.opt.ClusterSlots(ctx) |
|
if err != nil { |
|
return nil, err |
|
} |
|
return newClusterState(c.nodes, slots, "") |
|
} |
|
|
|
addrs, err := c.nodes.Addrs() |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
var firstErr error |
|
|
|
for _, idx := range rand.Perm(len(addrs)) { |
|
addr := addrs[idx] |
|
|
|
node, err := c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
if firstErr == nil { |
|
firstErr = err |
|
} |
|
continue |
|
} |
|
|
|
slots, err := node.Client.ClusterSlots(ctx).Result() |
|
if err != nil { |
|
if firstErr == nil { |
|
firstErr = err |
|
} |
|
continue |
|
} |
|
|
|
return newClusterState(c.nodes, slots, node.Client.opt.Addr) |
|
} |
|
|
|
/* |
|
* No node is connectable. It's possible that all nodes' IP has changed. |
|
* Clear activeAddrs to let client be able to re-connect using the initial |
|
* setting of the addresses (e.g. [redis-cluster-0:6379, redis-cluster-1:6379]), |
|
* which might have chance to resolve domain name and get updated IP address. |
|
*/ |
|
c.nodes.mu.Lock() |
|
c.nodes.activeAddrs = nil |
|
c.nodes.mu.Unlock() |
|
|
|
return nil, firstErr |
|
} |
|
|
|
// reaper closes idle connections to the cluster. |
|
func (c *ClusterClient) reaper(idleCheckFrequency time.Duration) { |
|
ticker := time.NewTicker(idleCheckFrequency) |
|
defer ticker.Stop() |
|
|
|
for range ticker.C { |
|
nodes, err := c.nodes.All() |
|
if err != nil { |
|
break |
|
} |
|
|
|
for _, node := range nodes { |
|
_, err := node.Client.connPool.(*pool.ConnPool).ReapStaleConns() |
|
if err != nil { |
|
internal.Logger.Printf(c.Context(), "ReapStaleConns failed: %s", err) |
|
} |
|
} |
|
} |
|
} |
|
|
|
func (c *ClusterClient) Pipeline() Pipeliner { |
|
pipe := Pipeline{ |
|
ctx: c.ctx, |
|
exec: c.processPipeline, |
|
} |
|
pipe.init() |
|
return &pipe |
|
} |
|
|
|
func (c *ClusterClient) Pipelined(ctx context.Context, fn func(Pipeliner) error) ([]Cmder, error) { |
|
return c.Pipeline().Pipelined(ctx, fn) |
|
} |
|
|
|
func (c *ClusterClient) processPipeline(ctx context.Context, cmds []Cmder) error { |
|
return c.hooks.processPipeline(ctx, cmds, c._processPipeline) |
|
} |
|
|
|
func (c *ClusterClient) _processPipeline(ctx context.Context, cmds []Cmder) error { |
|
cmdsMap := newCmdsMap() |
|
err := c.mapCmdsByNode(ctx, cmdsMap, cmds) |
|
if err != nil { |
|
setCmdsErr(cmds, err) |
|
return err |
|
} |
|
|
|
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ { |
|
if attempt > 0 { |
|
if err := internal.Sleep(ctx, c.retryBackoff(attempt)); err != nil { |
|
setCmdsErr(cmds, err) |
|
return err |
|
} |
|
} |
|
|
|
failedCmds := newCmdsMap() |
|
var wg sync.WaitGroup |
|
|
|
for node, cmds := range cmdsMap.m { |
|
wg.Add(1) |
|
go func(node *clusterNode, cmds []Cmder) { |
|
defer wg.Done() |
|
|
|
err := c._processPipelineNode(ctx, node, cmds, failedCmds) |
|
if err == nil { |
|
return |
|
} |
|
if attempt < c.opt.MaxRedirects { |
|
if err := c.mapCmdsByNode(ctx, failedCmds, cmds); err != nil { |
|
setCmdsErr(cmds, err) |
|
} |
|
} else { |
|
setCmdsErr(cmds, err) |
|
} |
|
}(node, cmds) |
|
} |
|
|
|
wg.Wait() |
|
if len(failedCmds.m) == 0 { |
|
break |
|
} |
|
cmdsMap = failedCmds |
|
} |
|
|
|
return cmdsFirstErr(cmds) |
|
} |
|
|
|
func (c *ClusterClient) mapCmdsByNode(ctx context.Context, cmdsMap *cmdsMap, cmds []Cmder) error { |
|
state, err := c.state.Get(ctx) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
if c.opt.ReadOnly && c.cmdsAreReadOnly(cmds) { |
|
for _, cmd := range cmds { |
|
slot := c.cmdSlot(cmd) |
|
node, err := c.slotReadOnlyNode(state, slot) |
|
if err != nil { |
|
return err |
|
} |
|
cmdsMap.Add(node, cmd) |
|
} |
|
return nil |
|
} |
|
|
|
for _, cmd := range cmds { |
|
slot := c.cmdSlot(cmd) |
|
node, err := state.slotMasterNode(slot) |
|
if err != nil { |
|
return err |
|
} |
|
cmdsMap.Add(node, cmd) |
|
} |
|
return nil |
|
} |
|
|
|
func (c *ClusterClient) cmdsAreReadOnly(cmds []Cmder) bool { |
|
for _, cmd := range cmds { |
|
cmdInfo := c.cmdInfo(cmd.Name()) |
|
if cmdInfo == nil || !cmdInfo.ReadOnly { |
|
return false |
|
} |
|
} |
|
return true |
|
} |
|
|
|
func (c *ClusterClient) _processPipelineNode( |
|
ctx context.Context, node *clusterNode, cmds []Cmder, failedCmds *cmdsMap, |
|
) error { |
|
return node.Client.hooks.processPipeline(ctx, cmds, func(ctx context.Context, cmds []Cmder) error { |
|
return node.Client.withConn(ctx, func(ctx context.Context, cn *pool.Conn) error { |
|
err := cn.WithWriter(ctx, c.opt.WriteTimeout, func(wr *proto.Writer) error { |
|
return writeCmds(wr, cmds) |
|
}) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
return cn.WithReader(ctx, c.opt.ReadTimeout, func(rd *proto.Reader) error { |
|
return c.pipelineReadCmds(ctx, node, rd, cmds, failedCmds) |
|
}) |
|
}) |
|
}) |
|
} |
|
|
|
func (c *ClusterClient) pipelineReadCmds( |
|
ctx context.Context, |
|
node *clusterNode, |
|
rd *proto.Reader, |
|
cmds []Cmder, |
|
failedCmds *cmdsMap, |
|
) error { |
|
for _, cmd := range cmds { |
|
err := cmd.readReply(rd) |
|
cmd.SetErr(err) |
|
|
|
if err == nil { |
|
continue |
|
} |
|
|
|
if c.checkMovedErr(ctx, cmd, err, failedCmds) { |
|
continue |
|
} |
|
|
|
if c.opt.ReadOnly && isLoadingError(err) { |
|
node.MarkAsFailing() |
|
return err |
|
} |
|
if isRedisError(err) { |
|
continue |
|
} |
|
return err |
|
} |
|
return nil |
|
} |
|
|
|
func (c *ClusterClient) checkMovedErr( |
|
ctx context.Context, cmd Cmder, err error, failedCmds *cmdsMap, |
|
) bool { |
|
moved, ask, addr := isMovedError(err) |
|
if !moved && !ask { |
|
return false |
|
} |
|
|
|
node, err := c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
return false |
|
} |
|
|
|
if moved { |
|
c.state.LazyReload() |
|
failedCmds.Add(node, cmd) |
|
return true |
|
} |
|
|
|
if ask { |
|
failedCmds.Add(node, NewCmd(ctx, "asking"), cmd) |
|
return true |
|
} |
|
|
|
panic("not reached") |
|
} |
|
|
|
// TxPipeline acts like Pipeline, but wraps queued commands with MULTI/EXEC. |
|
func (c *ClusterClient) TxPipeline() Pipeliner { |
|
pipe := Pipeline{ |
|
ctx: c.ctx, |
|
exec: c.processTxPipeline, |
|
} |
|
pipe.init() |
|
return &pipe |
|
} |
|
|
|
func (c *ClusterClient) TxPipelined(ctx context.Context, fn func(Pipeliner) error) ([]Cmder, error) { |
|
return c.TxPipeline().Pipelined(ctx, fn) |
|
} |
|
|
|
func (c *ClusterClient) processTxPipeline(ctx context.Context, cmds []Cmder) error { |
|
return c.hooks.processTxPipeline(ctx, cmds, c._processTxPipeline) |
|
} |
|
|
|
func (c *ClusterClient) _processTxPipeline(ctx context.Context, cmds []Cmder) error { |
|
// Trim multi .. exec. |
|
cmds = cmds[1 : len(cmds)-1] |
|
|
|
state, err := c.state.Get(ctx) |
|
if err != nil { |
|
setCmdsErr(cmds, err) |
|
return err |
|
} |
|
|
|
cmdsMap := c.mapCmdsBySlot(cmds) |
|
for slot, cmds := range cmdsMap { |
|
node, err := state.slotMasterNode(slot) |
|
if err != nil { |
|
setCmdsErr(cmds, err) |
|
continue |
|
} |
|
|
|
cmdsMap := map[*clusterNode][]Cmder{node: cmds} |
|
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ { |
|
if attempt > 0 { |
|
if err := internal.Sleep(ctx, c.retryBackoff(attempt)); err != nil { |
|
setCmdsErr(cmds, err) |
|
return err |
|
} |
|
} |
|
|
|
failedCmds := newCmdsMap() |
|
var wg sync.WaitGroup |
|
|
|
for node, cmds := range cmdsMap { |
|
wg.Add(1) |
|
go func(node *clusterNode, cmds []Cmder) { |
|
defer wg.Done() |
|
|
|
err := c._processTxPipelineNode(ctx, node, cmds, failedCmds) |
|
if err == nil { |
|
return |
|
} |
|
|
|
if attempt < c.opt.MaxRedirects { |
|
if err := c.mapCmdsByNode(ctx, failedCmds, cmds); err != nil { |
|
setCmdsErr(cmds, err) |
|
} |
|
} else { |
|
setCmdsErr(cmds, err) |
|
} |
|
}(node, cmds) |
|
} |
|
|
|
wg.Wait() |
|
if len(failedCmds.m) == 0 { |
|
break |
|
} |
|
cmdsMap = failedCmds.m |
|
} |
|
} |
|
|
|
return cmdsFirstErr(cmds) |
|
} |
|
|
|
func (c *ClusterClient) mapCmdsBySlot(cmds []Cmder) map[int][]Cmder { |
|
cmdsMap := make(map[int][]Cmder) |
|
for _, cmd := range cmds { |
|
slot := c.cmdSlot(cmd) |
|
cmdsMap[slot] = append(cmdsMap[slot], cmd) |
|
} |
|
return cmdsMap |
|
} |
|
|
|
func (c *ClusterClient) _processTxPipelineNode( |
|
ctx context.Context, node *clusterNode, cmds []Cmder, failedCmds *cmdsMap, |
|
) error { |
|
return node.Client.hooks.processTxPipeline(ctx, cmds, func(ctx context.Context, cmds []Cmder) error { |
|
return node.Client.withConn(ctx, func(ctx context.Context, cn *pool.Conn) error { |
|
err := cn.WithWriter(ctx, c.opt.WriteTimeout, func(wr *proto.Writer) error { |
|
return writeCmds(wr, cmds) |
|
}) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
return cn.WithReader(ctx, c.opt.ReadTimeout, func(rd *proto.Reader) error { |
|
statusCmd := cmds[0].(*StatusCmd) |
|
// Trim multi and exec. |
|
cmds = cmds[1 : len(cmds)-1] |
|
|
|
err := c.txPipelineReadQueued(ctx, rd, statusCmd, cmds, failedCmds) |
|
if err != nil { |
|
moved, ask, addr := isMovedError(err) |
|
if moved || ask { |
|
return c.cmdsMoved(ctx, cmds, moved, ask, addr, failedCmds) |
|
} |
|
return err |
|
} |
|
|
|
return pipelineReadCmds(rd, cmds) |
|
}) |
|
}) |
|
}) |
|
} |
|
|
|
func (c *ClusterClient) txPipelineReadQueued( |
|
ctx context.Context, |
|
rd *proto.Reader, |
|
statusCmd *StatusCmd, |
|
cmds []Cmder, |
|
failedCmds *cmdsMap, |
|
) error { |
|
// Parse queued replies. |
|
if err := statusCmd.readReply(rd); err != nil { |
|
return err |
|
} |
|
|
|
for _, cmd := range cmds { |
|
err := statusCmd.readReply(rd) |
|
if err == nil || c.checkMovedErr(ctx, cmd, err, failedCmds) || isRedisError(err) { |
|
continue |
|
} |
|
return err |
|
} |
|
|
|
// Parse number of replies. |
|
line, err := rd.ReadLine() |
|
if err != nil { |
|
if err == Nil { |
|
err = TxFailedErr |
|
} |
|
return err |
|
} |
|
|
|
switch line[0] { |
|
case proto.ErrorReply: |
|
return proto.ParseErrorReply(line) |
|
case proto.ArrayReply: |
|
// ok |
|
default: |
|
return fmt.Errorf("redis: expected '*', but got line %q", line) |
|
} |
|
|
|
return nil |
|
} |
|
|
|
func (c *ClusterClient) cmdsMoved( |
|
ctx context.Context, cmds []Cmder, |
|
moved, ask bool, |
|
addr string, |
|
failedCmds *cmdsMap, |
|
) error { |
|
node, err := c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
if moved { |
|
c.state.LazyReload() |
|
for _, cmd := range cmds { |
|
failedCmds.Add(node, cmd) |
|
} |
|
return nil |
|
} |
|
|
|
if ask { |
|
for _, cmd := range cmds { |
|
failedCmds.Add(node, NewCmd(ctx, "asking"), cmd) |
|
} |
|
return nil |
|
} |
|
|
|
return nil |
|
} |
|
|
|
func (c *ClusterClient) Watch(ctx context.Context, fn func(*Tx) error, keys ...string) error { |
|
if len(keys) == 0 { |
|
return fmt.Errorf("redis: Watch requires at least one key") |
|
} |
|
|
|
slot := hashtag.Slot(keys[0]) |
|
for _, key := range keys[1:] { |
|
if hashtag.Slot(key) != slot { |
|
err := fmt.Errorf("redis: Watch requires all keys to be in the same slot") |
|
return err |
|
} |
|
} |
|
|
|
node, err := c.slotMasterNode(ctx, slot) |
|
if err != nil { |
|
return err |
|
} |
|
|
|
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ { |
|
if attempt > 0 { |
|
if err := internal.Sleep(ctx, c.retryBackoff(attempt)); err != nil { |
|
return err |
|
} |
|
} |
|
|
|
err = node.Client.Watch(ctx, fn, keys...) |
|
if err == nil { |
|
break |
|
} |
|
|
|
moved, ask, addr := isMovedError(err) |
|
if moved || ask { |
|
node, err = c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
return err |
|
} |
|
continue |
|
} |
|
|
|
if isReadOnly := isReadOnlyError(err); isReadOnly || err == pool.ErrClosed { |
|
if isReadOnly { |
|
c.state.LazyReload() |
|
} |
|
node, err = c.slotMasterNode(ctx, slot) |
|
if err != nil { |
|
return err |
|
} |
|
continue |
|
} |
|
|
|
if shouldRetry(err, true) { |
|
continue |
|
} |
|
|
|
return err |
|
} |
|
|
|
return err |
|
} |
|
|
|
func (c *ClusterClient) pubSub() *PubSub { |
|
var node *clusterNode |
|
pubsub := &PubSub{ |
|
opt: c.opt.clientOptions(), |
|
|
|
newConn: func(ctx context.Context, channels []string) (*pool.Conn, error) { |
|
if node != nil { |
|
panic("node != nil") |
|
} |
|
|
|
var err error |
|
if len(channels) > 0 { |
|
slot := hashtag.Slot(channels[0]) |
|
node, err = c.slotMasterNode(ctx, slot) |
|
} else { |
|
node, err = c.nodes.Random() |
|
} |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
cn, err := node.Client.newConn(context.TODO()) |
|
if err != nil { |
|
node = nil |
|
|
|
return nil, err |
|
} |
|
|
|
return cn, nil |
|
}, |
|
closeConn: func(cn *pool.Conn) error { |
|
err := node.Client.connPool.CloseConn(cn) |
|
node = nil |
|
return err |
|
}, |
|
} |
|
pubsub.init() |
|
|
|
return pubsub |
|
} |
|
|
|
// Subscribe subscribes the client to the specified channels. |
|
// Channels can be omitted to create empty subscription. |
|
func (c *ClusterClient) Subscribe(ctx context.Context, channels ...string) *PubSub { |
|
pubsub := c.pubSub() |
|
if len(channels) > 0 { |
|
_ = pubsub.Subscribe(ctx, channels...) |
|
} |
|
return pubsub |
|
} |
|
|
|
// PSubscribe subscribes the client to the given patterns. |
|
// Patterns can be omitted to create empty subscription. |
|
func (c *ClusterClient) PSubscribe(ctx context.Context, channels ...string) *PubSub { |
|
pubsub := c.pubSub() |
|
if len(channels) > 0 { |
|
_ = pubsub.PSubscribe(ctx, channels...) |
|
} |
|
return pubsub |
|
} |
|
|
|
func (c *ClusterClient) retryBackoff(attempt int) time.Duration { |
|
return internal.RetryBackoff(attempt, c.opt.MinRetryBackoff, c.opt.MaxRetryBackoff) |
|
} |
|
|
|
func (c *ClusterClient) cmdsInfo(ctx context.Context) (map[string]*CommandInfo, error) { |
|
// Try 3 random nodes. |
|
const nodeLimit = 3 |
|
|
|
addrs, err := c.nodes.Addrs() |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
var firstErr error |
|
|
|
perm := rand.Perm(len(addrs)) |
|
if len(perm) > nodeLimit { |
|
perm = perm[:nodeLimit] |
|
} |
|
|
|
for _, idx := range perm { |
|
addr := addrs[idx] |
|
|
|
node, err := c.nodes.GetOrCreate(addr) |
|
if err != nil { |
|
if firstErr == nil { |
|
firstErr = err |
|
} |
|
continue |
|
} |
|
|
|
info, err := node.Client.Command(ctx).Result() |
|
if err == nil { |
|
return info, nil |
|
} |
|
if firstErr == nil { |
|
firstErr = err |
|
} |
|
} |
|
|
|
if firstErr == nil { |
|
panic("not reached") |
|
} |
|
return nil, firstErr |
|
} |
|
|
|
func (c *ClusterClient) cmdInfo(name string) *CommandInfo { |
|
cmdsInfo, err := c.cmdsInfoCache.Get(c.ctx) |
|
if err != nil { |
|
return nil |
|
} |
|
|
|
info := cmdsInfo[name] |
|
if info == nil { |
|
internal.Logger.Printf(c.Context(), "info for cmd=%s not found", name) |
|
} |
|
return info |
|
} |
|
|
|
func (c *ClusterClient) cmdSlot(cmd Cmder) int { |
|
args := cmd.Args() |
|
if args[0] == "cluster" && args[1] == "getkeysinslot" { |
|
return args[2].(int) |
|
} |
|
|
|
cmdInfo := c.cmdInfo(cmd.Name()) |
|
return cmdSlot(cmd, cmdFirstKeyPos(cmd, cmdInfo)) |
|
} |
|
|
|
func cmdSlot(cmd Cmder, pos int) int { |
|
if pos == 0 { |
|
return hashtag.RandomSlot() |
|
} |
|
firstKey := cmd.stringArg(pos) |
|
return hashtag.Slot(firstKey) |
|
} |
|
|
|
func (c *ClusterClient) cmdNode( |
|
ctx context.Context, |
|
cmdInfo *CommandInfo, |
|
slot int, |
|
) (*clusterNode, error) { |
|
state, err := c.state.Get(ctx) |
|
if err != nil { |
|
return nil, err |
|
} |
|
|
|
if c.opt.ReadOnly && cmdInfo != nil && cmdInfo.ReadOnly { |
|
return c.slotReadOnlyNode(state, slot) |
|
} |
|
return state.slotMasterNode(slot) |
|
} |
|
|
|
func (c *clusterClient) slotReadOnlyNode(state *clusterState, slot int) (*clusterNode, error) { |
|
if c.opt.RouteByLatency { |
|
return state.slotClosestNode(slot) |
|
} |
|
if c.opt.RouteRandomly { |
|
return state.slotRandomNode(slot) |
|
} |
|
return state.slotSlaveNode(slot) |
|
} |
|
|
|
func (c *ClusterClient) slotMasterNode(ctx context.Context, slot int) (*clusterNode, error) { |
|
state, err := c.state.Get(ctx) |
|
if err != nil { |
|
return nil, err |
|
} |
|
return state.slotMasterNode(slot) |
|
} |
|
|
|
// SlaveForKey gets a client for a replica node to run any command on it. |
|
// This is especially useful if we want to run a particular lua script which has |
|
// only read only commands on the replica. |
|
// This is because other redis commands generally have a flag that points that |
|
// they are read only and automatically run on the replica nodes |
|
// if ClusterOptions.ReadOnly flag is set to true. |
|
func (c *ClusterClient) SlaveForKey(ctx context.Context, key string) (*Client, error) { |
|
state, err := c.state.Get(ctx) |
|
if err != nil { |
|
return nil, err |
|
} |
|
slot := hashtag.Slot(key) |
|
node, err := c.slotReadOnlyNode(state, slot) |
|
if err != nil { |
|
return nil, err |
|
} |
|
return node.Client, err |
|
} |
|
|
|
// MasterForKey return a client to the master node for a particular key. |
|
func (c *ClusterClient) MasterForKey(ctx context.Context, key string) (*Client, error) { |
|
slot := hashtag.Slot(key) |
|
node, err := c.slotMasterNode(ctx, slot) |
|
if err != nil { |
|
return nil, err |
|
} |
|
return node.Client, err |
|
} |
|
|
|
func appendUniqueNode(nodes []*clusterNode, node *clusterNode) []*clusterNode { |
|
for _, n := range nodes { |
|
if n == node { |
|
return nodes |
|
} |
|
} |
|
return append(nodes, node) |
|
} |
|
|
|
func appendIfNotExists(ss []string, es ...string) []string { |
|
loop: |
|
for _, e := range es { |
|
for _, s := range ss { |
|
if s == e { |
|
continue loop |
|
} |
|
} |
|
ss = append(ss, e) |
|
} |
|
return ss |
|
} |
|
|
|
//------------------------------------------------------------------------------ |
|
|
|
type cmdsMap struct { |
|
mu sync.Mutex |
|
m map[*clusterNode][]Cmder |
|
} |
|
|
|
func newCmdsMap() *cmdsMap { |
|
return &cmdsMap{ |
|
m: make(map[*clusterNode][]Cmder), |
|
} |
|
} |
|
|
|
func (m *cmdsMap) Add(node *clusterNode, cmds ...Cmder) { |
|
m.mu.Lock() |
|
m.m[node] = append(m.m[node], cmds...) |
|
m.mu.Unlock() |
|
}
|
|
|