package redis import ( "math/rand" "sync" "sync/atomic" "time" "gopkg.in/redis.v4/internal" "gopkg.in/redis.v4/internal/hashtag" "gopkg.in/redis.v4/internal/pool" ) type clusterNode struct { Addr string Latency int Client *Client } // ClusterClient is a Redis Cluster client representing a pool of zero // or more underlying connections. It's safe for concurrent use by // multiple goroutines. type ClusterClient struct { commandable opt *ClusterOptions mu sync.RWMutex addrs []string nodes map[string]*clusterNode slots [][]*clusterNode closed bool cmdsInfo map[string]*CommandInfo cmdsInfoOnce *sync.Once // Reports where slots reloading is in progress. reloading uint32 } // NewClusterClient returns a Redis Cluster client as described in // http://redis.io/topics/cluster-spec. func NewClusterClient(opt *ClusterOptions) *ClusterClient { if opt.RouteByLatency { opt.ReadOnly = true } client := &ClusterClient{ opt: opt, nodes: make(map[string]*clusterNode), cmdsInfoOnce: new(sync.Once), } client.commandable.process = client.process for _, addr := range opt.Addrs { _ = client.nodeByAddr(addr) } client.reloadSlots() return client } func (c *ClusterClient) cmdInfo(name string) *CommandInfo { c.cmdsInfoOnce.Do(func() { for _, node := range c.nodes { cmdsInfo, err := node.Client.Command().Result() if err == nil { c.cmdsInfo = cmdsInfo return } } c.cmdsInfoOnce = &sync.Once{} }) return c.cmdsInfo[name] } func (c *ClusterClient) getNodes() map[string]*clusterNode { c.mu.RLock() var nodes map[string]*clusterNode if !c.closed { nodes = make(map[string]*clusterNode, len(c.nodes)) for addr, node := range c.nodes { nodes[addr] = node } } c.mu.RUnlock() return nodes } func (c *ClusterClient) Watch(fn func(*Tx) error, keys ...string) error { node := c.slotMasterNode(hashtag.Slot(keys[0])) return node.Client.Watch(fn, keys...) } // PoolStats returns accumulated connection pool stats. func (c *ClusterClient) PoolStats() *PoolStats { acc := PoolStats{} for _, node := range c.getNodes() { s := node.Client.connPool.Stats() acc.Requests += s.Requests acc.Hits += s.Hits acc.Timeouts += s.Timeouts acc.TotalConns += s.TotalConns acc.FreeConns += s.FreeConns } return &acc } // Close closes the cluster client, releasing any open resources. // // It is rare to Close a ClusterClient, as the ClusterClient is meant // to be long-lived and shared between many goroutines. func (c *ClusterClient) Close() error { c.mu.Lock() if !c.closed { c.closeClients() c.addrs = nil c.nodes = nil c.slots = nil c.cmdsInfo = nil } c.closed = true c.mu.Unlock() return nil } func (c *ClusterClient) nodeByAddr(addr string) *clusterNode { c.mu.RLock() node, ok := c.nodes[addr] c.mu.RUnlock() if ok { return node } c.mu.Lock() if !c.closed { node, ok = c.nodes[addr] if !ok { node = c.newNode(addr) c.nodes[addr] = node c.addrs = append(c.addrs, node.Addr) } } c.mu.Unlock() return node } func (c *ClusterClient) newNode(addr string) *clusterNode { opt := c.opt.clientOptions() opt.Addr = addr return &clusterNode{ Addr: addr, Client: NewClient(opt), } } func (c *ClusterClient) slotNodes(slot int) []*clusterNode { c.mu.RLock() nodes := c.slots[slot] c.mu.RUnlock() return nodes } // randomNode returns random live node. func (c *ClusterClient) randomNode() *clusterNode { var node *clusterNode for i := 0; i < 10; i++ { c.mu.RLock() addrs := c.addrs c.mu.RUnlock() if len(addrs) == 0 { return nil } n := rand.Intn(len(addrs)) node = c.nodeByAddr(addrs[n]) if node.Client.ClusterInfo().Err() == nil { return node } } return node } func (c *ClusterClient) slotMasterNode(slot int) *clusterNode { nodes := c.slotNodes(slot) if len(nodes) == 0 { return c.randomNode() } return nodes[0] } func (c *ClusterClient) slotSlaveNode(slot int) *clusterNode { nodes := c.slotNodes(slot) switch len(nodes) { case 0: return c.randomNode() case 1: return nodes[0] case 2: return nodes[1] default: n := rand.Intn(len(nodes)-1) + 1 return nodes[n] } } func (c *ClusterClient) slotClosestNode(slot int) *clusterNode { nodes := c.slotNodes(slot) var node *clusterNode for _, n := range nodes { if node == nil || n.Latency < node.Latency { node = n } } return node } func (c *ClusterClient) cmdSlotAndNode(cmd Cmder) (int, *clusterNode) { cmdInfo := c.cmdInfo(cmd.arg(0)) if cmdInfo == nil { return 0, c.randomNode() } if cmdInfo.FirstKeyPos == -1 { return 0, c.randomNode() } firstKey := cmd.arg(int(cmdInfo.FirstKeyPos)) slot := hashtag.Slot(firstKey) if cmdInfo.ReadOnly && c.opt.ReadOnly { if c.opt.RouteByLatency { return slot, c.slotClosestNode(slot) } return slot, c.slotSlaveNode(slot) } return slot, c.slotMasterNode(slot) } func (c *ClusterClient) process(cmd Cmder) { var ask bool slot, node := c.cmdSlotAndNode(cmd) for attempt := 0; attempt <= c.opt.getMaxRedirects(); attempt++ { if attempt > 0 { cmd.reset() } if node == nil { cmd.setErr(pool.ErrClosed) return } if ask { pipe := node.Client.Pipeline() pipe.Process(NewCmd("ASKING")) pipe.Process(cmd) _, _ = pipe.Exec() pipe.Close() ask = false } else { node.Client.Process(cmd) } // If there is no (real) error, we are done! err := cmd.Err() if err == nil { return } // On network errors try random node. if shouldRetry(err) { node = c.randomNode() continue } var moved bool var addr string moved, ask, addr = isMovedError(err) if moved || ask { if moved && c.slotMasterNode(slot).Addr != addr { c.lazyReloadSlots() } node = c.nodeByAddr(addr) continue } break } } // closeClients closes all clients and returns the first error if there are any. func (c *ClusterClient) closeClients() error { var retErr error for _, node := range c.nodes { if err := node.Client.Close(); err != nil && retErr == nil { retErr = err } } return retErr } func (c *ClusterClient) setSlots(cs []ClusterSlot) { slots := make([][]*clusterNode, hashtag.SlotNumber) for i := 0; i < hashtag.SlotNumber; i++ { slots[i] = nil } for _, s := range cs { var nodes []*clusterNode for _, n := range s.Nodes { nodes = append(nodes, c.nodeByAddr(n.Addr)) } for i := s.Start; i <= s.End; i++ { slots[i] = nodes } } c.mu.Lock() if !c.closed { c.slots = slots } c.mu.Unlock() } func (c *ClusterClient) setNodesLatency() { nodes := c.getNodes() for _, node := range nodes { var latency int for i := 0; i < 10; i++ { t1 := time.Now() node.Client.Ping() latency += int(time.Since(t1) / time.Millisecond) } node.Latency = latency } } func (c *ClusterClient) reloadSlots() { defer atomic.StoreUint32(&c.reloading, 0) node := c.randomNode() if node == nil { return } slots, err := node.Client.ClusterSlots().Result() if err != nil { internal.Logf("ClusterSlots on addr=%q failed: %s", node.Addr, err) return } c.setSlots(slots) if c.opt.RouteByLatency { c.setNodesLatency() } } func (c *ClusterClient) lazyReloadSlots() { if !atomic.CompareAndSwapUint32(&c.reloading, 0, 1) { return } go c.reloadSlots() } // reaper closes idle connections to the cluster. func (c *ClusterClient) reaper(frequency time.Duration) { ticker := time.NewTicker(frequency) defer ticker.Stop() for _ = range ticker.C { nodes := c.getNodes() if nodes == nil { break } var n int for _, node := range nodes { nn, err := node.Client.connPool.(*pool.ConnPool).ReapStaleConns() if err != nil { internal.Logf("ReapStaleConns failed: %s", err) } else { n += nn } } s := c.PoolStats() internal.Logf( "reaper: removed %d stale conns (TotalConns=%d FreeConns=%d Requests=%d Hits=%d Timeouts=%d)", n, s.TotalConns, s.FreeConns, s.Requests, s.Hits, s.Timeouts, ) } } func (c *ClusterClient) Pipeline() *Pipeline { pipe := &Pipeline{ exec: c.pipelineExec, } pipe.commandable.process = pipe.process return pipe } func (c *ClusterClient) Pipelined(fn func(*Pipeline) error) ([]Cmder, error) { return c.Pipeline().pipelined(fn) } func (c *ClusterClient) pipelineExec(cmds []Cmder) error { var retErr error cmdsMap := make(map[*clusterNode][]Cmder) for _, cmd := range cmds { _, node := c.cmdSlotAndNode(cmd) cmdsMap[node] = append(cmdsMap[node], cmd) } for attempt := 0; attempt <= c.opt.getMaxRedirects(); attempt++ { failedCmds := make(map[*clusterNode][]Cmder) for node, cmds := range cmdsMap { if node == nil { node = c.randomNode() } cn, err := node.Client.conn() if err != nil { setCmdsErr(cmds, err) retErr = err continue } failedCmds, err = c.execClusterCmds(cn, cmds, failedCmds) if err != nil { retErr = err } node.Client.putConn(cn, err, false) } cmdsMap = failedCmds } return retErr } func (c *ClusterClient) execClusterCmds( cn *pool.Conn, cmds []Cmder, failedCmds map[*clusterNode][]Cmder, ) (map[*clusterNode][]Cmder, error) { if err := writeCmd(cn, cmds...); err != nil { setCmdsErr(cmds, err) return failedCmds, err } var firstCmdErr error for i, cmd := range cmds { err := cmd.readReply(cn) if err == nil { continue } if isNetworkError(err) { cmd.reset() failedCmds[nil] = append(failedCmds[nil], cmds[i:]...) break } else if moved, ask, addr := isMovedError(err); moved { c.lazyReloadSlots() cmd.reset() node := c.nodeByAddr(addr) failedCmds[node] = append(failedCmds[node], cmd) } else if ask { cmd.reset() node := c.nodeByAddr(addr) failedCmds[node] = append(failedCmds[node], NewCmd("ASKING"), cmd) } else if firstCmdErr == nil { firstCmdErr = err } } return failedCmds, firstCmdErr } //------------------------------------------------------------------------------ // ClusterOptions are used to configure a cluster client and should be // passed to NewClusterClient. type ClusterOptions struct { // A seed list of host:port addresses of cluster nodes. Addrs []string // The maximum number of retries before giving up. Command is retried // on network errors and MOVED/ASK redirects. // Default is 16. MaxRedirects int // Enables read queries for a connection to a Redis Cluster slave node. ReadOnly bool // Enables routing read-only queries to the closest master or slave node. RouteByLatency bool // Following options are copied from Options struct. Password string DialTimeout time.Duration ReadTimeout time.Duration WriteTimeout time.Duration // PoolSize applies per cluster node and not for the whole cluster. PoolSize int PoolTimeout time.Duration IdleTimeout time.Duration IdleCheckFrequency time.Duration } func (opt *ClusterOptions) getMaxRedirects() int { if opt.MaxRedirects == -1 { return 0 } if opt.MaxRedirects == 0 { return 16 } return opt.MaxRedirects } func (opt *ClusterOptions) clientOptions() *Options { return &Options{ Password: opt.Password, ReadOnly: opt.ReadOnly, DialTimeout: opt.DialTimeout, ReadTimeout: opt.ReadTimeout, WriteTimeout: opt.WriteTimeout, PoolSize: opt.PoolSize, PoolTimeout: opt.PoolTimeout, IdleTimeout: opt.IdleTimeout, // IdleCheckFrequency is not copied to disable reaper } }