redis/cluster.go

544 lines
11 KiB
Go
Raw Normal View History

2015-01-24 15:12:48 +03:00
package redis
import (
"math/rand"
"sync"
"sync/atomic"
"time"
2015-12-30 16:53:45 +03:00
2016-04-09 14:52:01 +03:00
"gopkg.in/redis.v4/internal"
"gopkg.in/redis.v4/internal/hashtag"
"gopkg.in/redis.v4/internal/pool"
2015-01-24 15:12:48 +03:00
)
type clusterNode struct {
Addr string
Latency int
Client *Client
}
2015-09-12 09:36:03 +03:00
// ClusterClient is a Redis Cluster client representing a pool of zero
// or more underlying connections. It's safe for concurrent use by
// multiple goroutines.
2015-01-24 15:12:48 +03:00
type ClusterClient struct {
cmdable
2015-01-24 15:12:48 +03:00
2016-03-17 19:00:47 +03:00
opt *ClusterOptions
mu sync.RWMutex
addrs []string
nodes map[string]*clusterNode
slots [][]*clusterNode
closed bool
2015-03-30 17:10:53 +03:00
cmdsInfo map[string]*CommandInfo
cmdsInfoOnce *sync.Once
2015-01-24 15:12:48 +03:00
2015-05-01 10:42:58 +03:00
// Reports where slots reloading is in progress.
reloading uint32
2015-01-24 15:12:48 +03:00
}
2015-09-12 09:36:03 +03:00
// NewClusterClient returns a Redis Cluster client as described in
// http://redis.io/topics/cluster-spec.
2015-04-04 16:46:57 +03:00
func NewClusterClient(opt *ClusterOptions) *ClusterClient {
2016-06-05 14:10:30 +03:00
opt.init()
2015-01-24 15:12:48 +03:00
client := &ClusterClient{
opt: opt,
nodes: make(map[string]*clusterNode),
cmdsInfoOnce: new(sync.Once),
2015-01-24 15:12:48 +03:00
}
client.cmdable.process = client.Process
for _, addr := range opt.Addrs {
_ = client.nodeByAddr(addr)
}
2015-05-01 10:42:58 +03:00
client.reloadSlots()
2015-04-04 16:46:57 +03:00
return client
2015-01-24 15:12:48 +03:00
}
func (c *ClusterClient) cmdInfo(name string) *CommandInfo {
c.cmdsInfoOnce.Do(func() {
for _, node := range c.nodes {
cmdsInfo, err := node.Client.Command().Result()
if err == nil {
c.cmdsInfo = cmdsInfo
return
}
}
c.cmdsInfoOnce = &sync.Once{}
})
return c.cmdsInfo[name]
2016-03-17 19:00:47 +03:00
}
func (c *ClusterClient) getNodes() map[string]*clusterNode {
c.mu.RLock()
var nodes map[string]*clusterNode
if !c.closed {
nodes = make(map[string]*clusterNode, len(c.nodes))
for addr, node := range c.nodes {
nodes[addr] = node
}
2015-12-16 17:11:52 +03:00
}
c.mu.RUnlock()
return nodes
}
func (c *ClusterClient) Watch(fn func(*Tx) error, keys ...string) error {
node := c.slotMasterNode(hashtag.Slot(keys[0]))
return node.Client.Watch(fn, keys...)
2015-12-16 17:11:52 +03:00
}
2016-01-25 16:57:09 +03:00
// PoolStats returns accumulated connection pool stats.
2016-01-19 19:36:40 +03:00
func (c *ClusterClient) PoolStats() *PoolStats {
acc := PoolStats{}
for _, node := range c.getNodes() {
s := node.Client.connPool.Stats()
2016-03-17 19:00:47 +03:00
acc.Requests += s.Requests
acc.Hits += s.Hits
acc.Timeouts += s.Timeouts
acc.TotalConns += s.TotalConns
acc.FreeConns += s.FreeConns
2016-01-19 19:36:40 +03:00
}
return &acc
}
// Close closes the cluster client, releasing any open resources.
//
2015-09-12 09:36:03 +03:00
// It is rare to Close a ClusterClient, as the ClusterClient is meant
// to be long-lived and shared between many goroutines.
2015-01-24 15:12:48 +03:00
func (c *ClusterClient) Close() error {
c.mu.Lock()
if !c.closed {
c.closeClients()
c.addrs = nil
c.nodes = nil
c.slots = nil
c.cmdsInfo = nil
}
c.closed = true
c.mu.Unlock()
2015-04-04 16:46:57 +03:00
return nil
2015-01-24 15:12:48 +03:00
}
func (c *ClusterClient) nodeByAddr(addr string) *clusterNode {
c.mu.RLock()
node, ok := c.nodes[addr]
c.mu.RUnlock()
if ok {
return node
2016-03-17 19:00:47 +03:00
}
c.mu.Lock()
if !c.closed {
node, ok = c.nodes[addr]
if !ok {
node = c.newNode(addr)
c.nodes[addr] = node
c.addrs = append(c.addrs, node.Addr)
}
2015-03-18 13:41:24 +03:00
}
c.mu.Unlock()
2015-03-18 13:41:24 +03:00
return node
}
2015-01-24 15:12:48 +03:00
func (c *ClusterClient) newNode(addr string) *clusterNode {
opt := c.opt.clientOptions()
opt.Addr = addr
return &clusterNode{
Addr: addr,
Client: NewClient(opt),
2015-01-24 15:12:48 +03:00
}
}
2015-04-04 16:46:57 +03:00
func (c *ClusterClient) slotNodes(slot int) []*clusterNode {
c.mu.RLock()
nodes := c.slots[slot]
c.mu.RUnlock()
return nodes
2015-03-18 13:41:24 +03:00
}
// randomNode returns random live node.
func (c *ClusterClient) randomNode() *clusterNode {
var node *clusterNode
for i := 0; i < 10; i++ {
c.mu.RLock()
addrs := c.addrs
c.mu.RUnlock()
if len(addrs) == 0 {
return nil
}
n := rand.Intn(len(addrs))
node = c.nodeByAddr(addrs[n])
if node.Client.ClusterInfo().Err() == nil {
return node
}
}
return node
2015-01-24 15:12:48 +03:00
}
func (c *ClusterClient) slotMasterNode(slot int) *clusterNode {
nodes := c.slotNodes(slot)
if len(nodes) == 0 {
return c.randomNode()
}
return nodes[0]
}
func (c *ClusterClient) slotSlaveNode(slot int) *clusterNode {
nodes := c.slotNodes(slot)
switch len(nodes) {
case 0:
return c.randomNode()
case 1:
return nodes[0]
case 2:
return nodes[1]
default:
n := rand.Intn(len(nodes)-1) + 1
return nodes[n]
}
}
func (c *ClusterClient) slotClosestNode(slot int) *clusterNode {
nodes := c.slotNodes(slot)
var node *clusterNode
for _, n := range nodes {
if node == nil || n.Latency < node.Latency {
node = n
2015-04-04 16:46:57 +03:00
}
}
return node
2015-04-04 16:46:57 +03:00
}
func (c *ClusterClient) cmdSlotAndNode(cmd Cmder) (int, *clusterNode) {
cmdInfo := c.cmdInfo(cmd.arg(0))
if cmdInfo == nil {
return 0, c.randomNode()
}
2015-01-24 15:12:48 +03:00
if cmdInfo.FirstKeyPos == -1 {
return 0, c.randomNode()
}
2015-01-24 15:12:48 +03:00
firstKey := cmd.arg(int(cmdInfo.FirstKeyPos))
slot := hashtag.Slot(firstKey)
if cmdInfo.ReadOnly && c.opt.ReadOnly {
if c.opt.RouteByLatency {
return slot, c.slotClosestNode(slot)
}
return slot, c.slotSlaveNode(slot)
2015-04-04 16:46:57 +03:00
}
return slot, c.slotMasterNode(slot)
}
func (c *ClusterClient) Process(cmd Cmder) {
var ask bool
slot, node := c.cmdSlotAndNode(cmd)
2015-04-04 16:46:57 +03:00
2016-06-05 14:10:30 +03:00
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ {
if attempt > 0 {
cmd.reset()
}
if node == nil {
cmd.setErr(pool.ErrClosed)
return
}
2015-01-24 15:12:48 +03:00
if ask {
pipe := node.Client.Pipeline()
2015-01-24 15:12:48 +03:00
pipe.Process(NewCmd("ASKING"))
pipe.Process(cmd)
_, _ = pipe.Exec()
pipe.Close()
2015-01-24 15:12:48 +03:00
ask = false
} else {
node.Client.Process(cmd)
2015-01-24 15:12:48 +03:00
}
// If there is no (real) error, we are done!
err := cmd.Err()
2016-03-19 17:33:14 +03:00
if err == nil {
2015-01-24 15:12:48 +03:00
return
}
2015-04-07 12:30:06 +03:00
// On network errors try random node.
2016-03-19 17:33:14 +03:00
if shouldRetry(err) {
node = c.randomNode()
2015-01-24 15:12:48 +03:00
continue
}
2015-03-18 13:41:24 +03:00
var moved bool
var addr string
moved, ask, addr = isMovedError(err)
if moved || ask {
if moved && c.slotMasterNode(slot).Addr != addr {
2015-05-01 10:42:58 +03:00
c.lazyReloadSlots()
2015-03-18 13:41:24 +03:00
}
node = c.nodeByAddr(addr)
2015-03-18 13:41:24 +03:00
continue
2015-01-24 15:12:48 +03:00
}
2015-03-18 13:41:24 +03:00
break
2015-01-24 15:12:48 +03:00
}
}
// closeClients closes all clients and returns the first error if there are any.
func (c *ClusterClient) closeClients() error {
var retErr error
for _, node := range c.nodes {
if err := node.Client.Close(); err != nil && retErr == nil {
2015-12-22 12:44:49 +03:00
retErr = err
2015-04-04 16:46:57 +03:00
}
2015-01-24 15:12:48 +03:00
}
2015-12-22 12:44:49 +03:00
return retErr
2015-04-04 16:46:57 +03:00
}
2015-01-24 15:12:48 +03:00
func (c *ClusterClient) setSlots(cs []ClusterSlot) {
slots := make([][]*clusterNode, hashtag.SlotNumber)
2015-12-30 16:53:45 +03:00
for i := 0; i < hashtag.SlotNumber; i++ {
slots[i] = nil
}
for _, s := range cs {
var nodes []*clusterNode
for _, n := range s.Nodes {
nodes = append(nodes, c.nodeByAddr(n.Addr))
2015-04-07 12:30:06 +03:00
}
for i := s.Start; i <= s.End; i++ {
slots[i] = nodes
}
}
c.mu.Lock()
if !c.closed {
c.slots = slots
2015-01-24 15:12:48 +03:00
}
c.mu.Unlock()
}
2015-01-24 15:12:48 +03:00
func (c *ClusterClient) setNodesLatency() {
nodes := c.getNodes()
for _, node := range nodes {
var latency int
for i := 0; i < 10; i++ {
t1 := time.Now()
node.Client.Ping()
latency += int(time.Since(t1) / time.Millisecond)
}
node.Latency = latency
}
2015-01-24 15:12:48 +03:00
}
2015-05-01 10:42:58 +03:00
func (c *ClusterClient) reloadSlots() {
defer atomic.StoreUint32(&c.reloading, 0)
2015-01-24 15:12:48 +03:00
node := c.randomNode()
if node == nil {
2015-05-01 10:42:58 +03:00
return
2015-01-24 15:12:48 +03:00
}
slots, err := node.Client.ClusterSlots().Result()
2015-04-04 16:46:57 +03:00
if err != nil {
internal.Logf("ClusterSlots on addr=%q failed: %s", node.Addr, err)
2015-05-01 10:42:58 +03:00
return
2015-04-04 16:46:57 +03:00
}
2015-04-04 16:46:57 +03:00
c.setSlots(slots)
if c.opt.RouteByLatency {
c.setNodesLatency()
}
2015-01-24 15:12:48 +03:00
}
2015-05-01 10:42:58 +03:00
func (c *ClusterClient) lazyReloadSlots() {
if !atomic.CompareAndSwapUint32(&c.reloading, 0, 1) {
return
}
go c.reloadSlots()
2015-01-24 15:12:48 +03:00
}
// reaper closes idle connections to the cluster.
2016-03-17 19:00:47 +03:00
func (c *ClusterClient) reaper(frequency time.Duration) {
ticker := time.NewTicker(frequency)
defer ticker.Stop()
2016-03-17 19:00:47 +03:00
for _ = range ticker.C {
nodes := c.getNodes()
if nodes == nil {
break
}
2016-03-17 19:00:47 +03:00
var n int
for _, node := range nodes {
nn, err := node.Client.connPool.(*pool.ConnPool).ReapStaleConns()
2016-03-17 19:00:47 +03:00
if err != nil {
2016-04-09 14:52:01 +03:00
internal.Logf("ReapStaleConns failed: %s", err)
2016-03-17 19:00:47 +03:00
} else {
n += nn
}
}
2016-03-17 19:00:47 +03:00
s := c.PoolStats()
2016-04-09 14:52:01 +03:00
internal.Logf(
2016-03-17 19:00:47 +03:00
"reaper: removed %d stale conns (TotalConns=%d FreeConns=%d Requests=%d Hits=%d Timeouts=%d)",
n, s.TotalConns, s.FreeConns, s.Requests, s.Hits, s.Timeouts,
)
}
}
func (c *ClusterClient) Pipeline() *Pipeline {
pipe := Pipeline{
exec: c.pipelineExec,
}
pipe.cmdable.process = pipe.Process
pipe.statefulCmdable.process = pipe.Process
return &pipe
}
func (c *ClusterClient) Pipelined(fn func(*Pipeline) error) ([]Cmder, error) {
return c.Pipeline().pipelined(fn)
}
func (c *ClusterClient) pipelineExec(cmds []Cmder) error {
var retErr error
cmdsMap := make(map[*clusterNode][]Cmder)
for _, cmd := range cmds {
_, node := c.cmdSlotAndNode(cmd)
cmdsMap[node] = append(cmdsMap[node], cmd)
}
2016-06-05 14:10:30 +03:00
for attempt := 0; attempt <= c.opt.MaxRedirects; attempt++ {
failedCmds := make(map[*clusterNode][]Cmder)
for node, cmds := range cmdsMap {
if node == nil {
node = c.randomNode()
}
cn, err := node.Client.conn()
if err != nil {
setCmdsErr(cmds, err)
retErr = err
continue
}
failedCmds, err = c.execClusterCmds(cn, cmds, failedCmds)
if err != nil {
retErr = err
}
node.Client.putConn(cn, err, false)
}
cmdsMap = failedCmds
}
return retErr
}
func (c *ClusterClient) execClusterCmds(
cn *pool.Conn, cmds []Cmder, failedCmds map[*clusterNode][]Cmder,
) (map[*clusterNode][]Cmder, error) {
if err := writeCmd(cn, cmds...); err != nil {
setCmdsErr(cmds, err)
return failedCmds, err
}
var firstCmdErr error
for i, cmd := range cmds {
err := cmd.readReply(cn)
if err == nil {
continue
}
if isNetworkError(err) {
cmd.reset()
failedCmds[nil] = append(failedCmds[nil], cmds[i:]...)
break
} else if moved, ask, addr := isMovedError(err); moved {
c.lazyReloadSlots()
cmd.reset()
node := c.nodeByAddr(addr)
failedCmds[node] = append(failedCmds[node], cmd)
} else if ask {
cmd.reset()
node := c.nodeByAddr(addr)
failedCmds[node] = append(failedCmds[node], NewCmd("ASKING"), cmd)
} else if firstCmdErr == nil {
firstCmdErr = err
}
}
return failedCmds, firstCmdErr
}
2015-01-24 15:12:48 +03:00
//------------------------------------------------------------------------------
// ClusterOptions are used to configure a cluster client and should be
// passed to NewClusterClient.
2015-01-24 15:12:48 +03:00
type ClusterOptions struct {
// A seed list of host:port addresses of cluster nodes.
2015-01-24 15:12:48 +03:00
Addrs []string
2016-04-06 13:13:03 +03:00
// The maximum number of retries before giving up. Command is retried
// on network errors and MOVED/ASK redirects.
// Default is 16.
MaxRedirects int
// Enables read queries for a connection to a Redis Cluster slave node.
ReadOnly bool
// Enables routing read-only queries to the closest master or slave node.
RouteByLatency bool
// Following options are copied from Options struct.
2015-01-24 15:12:48 +03:00
Password string
DialTimeout time.Duration
ReadTimeout time.Duration
WriteTimeout time.Duration
2015-01-24 15:12:48 +03:00
2016-01-25 16:57:09 +03:00
// PoolSize applies per cluster node and not for the whole cluster.
2016-03-17 19:00:47 +03:00
PoolSize int
PoolTimeout time.Duration
IdleTimeout time.Duration
IdleCheckFrequency time.Duration
2015-01-24 15:12:48 +03:00
}
2016-06-05 14:10:30 +03:00
func (opt *ClusterOptions) init() {
if opt.MaxRedirects == -1 {
2016-06-05 14:10:30 +03:00
opt.MaxRedirects = 0
} else if opt.MaxRedirects == 0 {
opt.MaxRedirects = 16
}
2016-06-05 14:10:30 +03:00
if opt.RouteByLatency {
opt.ReadOnly = true
2015-01-24 15:12:48 +03:00
}
}
func (opt *ClusterOptions) clientOptions() *Options {
return &Options{
Password: opt.Password,
ReadOnly: opt.ReadOnly,
2015-01-24 15:12:48 +03:00
DialTimeout: opt.DialTimeout,
2015-01-24 15:12:48 +03:00
ReadTimeout: opt.ReadTimeout,
WriteTimeout: opt.WriteTimeout,
PoolSize: opt.PoolSize,
PoolTimeout: opt.PoolTimeout,
2015-01-24 15:12:48 +03:00
IdleTimeout: opt.IdleTimeout,
2016-03-17 19:00:47 +03:00
// IdleCheckFrequency is not copied to disable reaper
2015-01-24 15:12:48 +03:00
}
}