Enable reaper on ClusterClient and add tests.

This commit is contained in:
Vladimir Mihailenco 2016-10-02 12:44:01 +00:00
parent a7d1d0b9ac
commit 7cbee9d337
11 changed files with 257 additions and 169 deletions

View File

@ -45,20 +45,25 @@ var _ Cmdable = (*ClusterClient)(nil)
// http://redis.io/topics/cluster-spec. // http://redis.io/topics/cluster-spec.
func NewClusterClient(opt *ClusterOptions) *ClusterClient { func NewClusterClient(opt *ClusterOptions) *ClusterClient {
opt.init() opt.init()
client := &ClusterClient{
c := &ClusterClient{
opt: opt, opt: opt,
nodes: make(map[string]*clusterNode), nodes: make(map[string]*clusterNode),
cmdsInfoOnce: new(sync.Once), cmdsInfoOnce: new(sync.Once),
} }
client.cmdable.process = client.Process c.cmdable.process = c.Process
for _, addr := range opt.Addrs { for _, addr := range opt.Addrs {
_, _ = client.nodeByAddr(addr) _, _ = c.nodeByAddr(addr)
} }
client.reloadSlots() c.reloadSlots()
return client if opt.IdleCheckFrequency > 0 {
go c.reaper(opt.IdleCheckFrequency)
}
return c
} }
func (c *ClusterClient) cmdInfo(name string) *CommandInfo { func (c *ClusterClient) cmdInfo(name string) *CommandInfo {
@ -333,11 +338,9 @@ func (c *ClusterClient) ForEachMaster(fn func(client *Client) error) error {
slots := c.slots slots := c.slots
c.mu.RUnlock() c.mu.RUnlock()
var retErr error
var mu sync.Mutex
var wg sync.WaitGroup var wg sync.WaitGroup
visited := make(map[*clusterNode]struct{}) visited := make(map[*clusterNode]struct{})
errCh := make(chan error, 1)
for _, nodes := range slots { for _, nodes := range slots {
if len(nodes) == 0 { if len(nodes) == 0 {
continue continue
@ -351,20 +354,24 @@ func (c *ClusterClient) ForEachMaster(fn func(client *Client) error) error {
wg.Add(1) wg.Add(1)
go func(node *clusterNode) { go func(node *clusterNode) {
defer wg.Done()
err := fn(node.Client) err := fn(node.Client)
if err != nil { if err != nil {
mu.Lock() select {
if retErr == nil { case errCh <- err:
retErr = err default:
} }
mu.Unlock()
} }
wg.Done()
}(master) }(master)
} }
wg.Wait() wg.Wait()
return retErr select {
case err := <-errCh:
return err
default:
return nil
}
} }
// closeClients closes all clients and returns the first error if there are any. // closeClients closes all clients and returns the first error if there are any.
@ -442,8 +449,8 @@ func (c *ClusterClient) setNodesLatency() {
} }
// reaper closes idle connections to the cluster. // reaper closes idle connections to the cluster.
func (c *ClusterClient) reaper(frequency time.Duration) { func (c *ClusterClient) reaper(idleCheckFrequency time.Duration) {
ticker := time.NewTicker(frequency) ticker := time.NewTicker(idleCheckFrequency)
defer ticker.Stop() defer ticker.Stop()
for _ = range ticker.C { for _ = range ticker.C {

View File

@ -2,7 +2,6 @@ package redis_test
import ( import (
"fmt" "fmt"
"math/rand"
"net" "net"
"strconv" "strconv"
"strings" "strings"
@ -50,17 +49,6 @@ func (s *clusterScenario) clusterClient(opt *redis.ClusterOptions) *redis.Cluste
for i, port := range s.ports { for i, port := range s.ports {
addrs[i] = net.JoinHostPort("127.0.0.1", port) addrs[i] = net.JoinHostPort("127.0.0.1", port)
} }
if opt == nil {
opt = &redis.ClusterOptions{
DialTimeout: 10 * time.Second,
ReadTimeout: 30 * time.Second,
WriteTimeout: 30 * time.Second,
PoolSize: 10,
PoolTimeout: 30 * time.Second,
IdleTimeout: time.Second,
IdleCheckFrequency: time.Second,
}
}
opt.Addrs = addrs opt.Addrs = addrs
return redis.NewClusterClient(opt) return redis.NewClusterClient(opt)
} }
@ -193,52 +181,12 @@ func stopCluster(scenario *clusterScenario) error {
//------------------------------------------------------------------------------ //------------------------------------------------------------------------------
var _ = Describe("Cluster", func() { var _ = Describe("ClusterClient", func() {
Describe("HashSlot", func() { var client *redis.ClusterClient
It("should calculate hash slots", func() {
tests := []struct {
key string
slot int
}{
{"123456789", 12739},
{"{}foo", 9500},
{"foo{}", 5542},
{"foo{}{bar}", 8363},
{"", 10503},
{"", 5176},
{string([]byte{83, 153, 134, 118, 229, 214, 244, 75, 140, 37, 215, 215}), 5463},
}
// Empty keys receive random slot.
rand.Seed(100)
for _, test := range tests {
Expect(hashtag.Slot(test.key)).To(Equal(test.slot), "for %s", test.key)
}
})
It("should extract keys from tags", func() {
tests := []struct {
one, two string
}{
{"foo{bar}", "bar"},
{"{foo}bar", "foo"},
{"{user1000}.following", "{user1000}.followers"},
{"foo{{bar}}zap", "{bar"},
{"foo{bar}{zap}", "bar"},
}
for _, test := range tests {
Expect(hashtag.Slot(test.one)).To(Equal(hashtag.Slot(test.two)), "for %s <-> %s", test.one, test.two)
}
})
})
Describe("Commands", func() {
describeClusterClient := func() {
It("should CLUSTER SLOTS", func() { It("should CLUSTER SLOTS", func() {
res, err := cluster.primary().ClusterSlots().Result() res, err := client.ClusterSlots().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(res).To(HaveLen(3)) Expect(res).To(HaveLen(3))
@ -251,73 +199,48 @@ var _ = Describe("Cluster", func() {
}) })
It("should CLUSTER NODES", func() { It("should CLUSTER NODES", func() {
res, err := cluster.primary().ClusterNodes().Result() res, err := client.ClusterNodes().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(len(res)).To(BeNumerically(">", 400)) Expect(len(res)).To(BeNumerically(">", 400))
}) })
It("should CLUSTER INFO", func() { It("should CLUSTER INFO", func() {
res, err := cluster.primary().ClusterInfo().Result() res, err := client.ClusterInfo().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(res).To(ContainSubstring("cluster_known_nodes:6")) Expect(res).To(ContainSubstring("cluster_known_nodes:6"))
}) })
It("should CLUSTER KEYSLOT", func() { It("should CLUSTER KEYSLOT", func() {
hashSlot, err := cluster.primary().ClusterKeySlot("somekey").Result() hashSlot, err := client.ClusterKeySlot("somekey").Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey")))) Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey"))))
}) })
It("should CLUSTER COUNT-FAILURE-REPORTS", func() { It("should CLUSTER COUNT-FAILURE-REPORTS", func() {
n, err := cluster.primary().ClusterCountFailureReports(cluster.nodeIds[0]).Result() n, err := client.ClusterCountFailureReports(cluster.nodeIds[0]).Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(n).To(Equal(int64(0))) Expect(n).To(Equal(int64(0)))
}) })
It("should CLUSTER COUNTKEYSINSLOT", func() { It("should CLUSTER COUNTKEYSINSLOT", func() {
n, err := cluster.primary().ClusterCountKeysInSlot(10).Result() n, err := client.ClusterCountKeysInSlot(10).Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(n).To(Equal(int64(0))) Expect(n).To(Equal(int64(0)))
}) })
It("should CLUSTER DELSLOTS", func() {
res, err := cluster.primary().ClusterDelSlotsRange(16000, 16384-1).Result()
Expect(err).NotTo(HaveOccurred())
Expect(res).To(Equal("OK"))
cluster.primary().ClusterAddSlotsRange(16000, 16384-1)
})
It("should CLUSTER SAVECONFIG", func() { It("should CLUSTER SAVECONFIG", func() {
res, err := cluster.primary().ClusterSaveConfig().Result() res, err := client.ClusterSaveConfig().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(res).To(Equal("OK")) Expect(res).To(Equal("OK"))
}) })
It("should CLUSTER SLAVES", func() { It("should CLUSTER SLAVES", func() {
nodesList, err := cluster.primary().ClusterSlaves(cluster.nodeIds[0]).Result() nodesList, err := client.ClusterSlaves(cluster.nodeIds[0]).Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(nodesList).Should(ContainElement(ContainSubstring("slave"))) Expect(nodesList).Should(ContainElement(ContainSubstring("slave")))
Expect(nodesList).Should(HaveLen(1)) Expect(nodesList).Should(HaveLen(1))
}) })
// It("should CLUSTER READONLY", func() {
// res, err := cluster.primary().ReadOnly().Result()
// Expect(err).NotTo(HaveOccurred())
// Expect(res).To(Equal("OK"))
// })
// It("should CLUSTER READWRITE", func() {
// res, err := cluster.primary().ReadWrite().Result()
// Expect(err).NotTo(HaveOccurred())
// Expect(res).To(Equal("OK"))
// })
})
})
var _ = Describe("ClusterClient", func() {
var client *redis.ClusterClient
describeClusterClient := func() {
It("should GET/SET/DEL", func() { It("should GET/SET/DEL", func() {
val, err := client.Get("A").Result() val, err := client.Get("A").Result()
Expect(err).To(Equal(redis.Nil)) Expect(err).To(Equal(redis.Nil))
@ -340,6 +263,18 @@ var _ = Describe("ClusterClient", func() {
Expect(client.PoolStats()).To(BeAssignableToTypeOf(&redis.PoolStats{})) Expect(client.PoolStats()).To(BeAssignableToTypeOf(&redis.PoolStats{}))
}) })
It("removes idle connections", func() {
stats := client.PoolStats()
Expect(stats.TotalConns).NotTo(BeZero())
Expect(stats.FreeConns).NotTo(BeZero())
time.Sleep(2 * time.Second)
stats = client.PoolStats()
Expect(stats.TotalConns).To(BeZero())
Expect(stats.FreeConns).To(BeZero())
})
It("follows redirects", func() { It("follows redirects", func() {
Expect(client.Set("A", "VALUE", 0).Err()).NotTo(HaveOccurred()) Expect(client.Set("A", "VALUE", 0).Err()).NotTo(HaveOccurred())
@ -352,9 +287,9 @@ var _ = Describe("ClusterClient", func() {
}) })
It("returns an error when there are no attempts left", func() { It("returns an error when there are no attempts left", func() {
client := cluster.clusterClient(&redis.ClusterOptions{ opt := redisClusterOptions()
MaxRedirects: -1, opt.MaxRedirects = -1
}) client := cluster.clusterClient(opt)
slot := hashtag.Slot("A") slot := hashtag.Slot("A")
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"})) Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
@ -483,7 +418,7 @@ var _ = Describe("ClusterClient", func() {
Describe("default ClusterClient", func() { Describe("default ClusterClient", func() {
BeforeEach(func() { BeforeEach(func() {
client = cluster.clusterClient(nil) client = cluster.clusterClient(redisClusterOptions())
_ = client.ForEachMaster(func(master *redis.Client) error { _ = client.ForEachMaster(func(master *redis.Client) error {
return master.FlushDb().Err() return master.FlushDb().Err()
@ -499,9 +434,9 @@ var _ = Describe("ClusterClient", func() {
Describe("ClusterClient with RouteByLatency", func() { Describe("ClusterClient with RouteByLatency", func() {
BeforeEach(func() { BeforeEach(func() {
client = cluster.clusterClient(&redis.ClusterOptions{ opt := redisClusterOptions()
RouteByLatency: true, opt.RouteByLatency = true
}) client = cluster.clusterClient(opt)
_ = client.ForEachMaster(func(master *redis.Client) error { _ = client.ForEachMaster(func(master *redis.Client) error {
return master.FlushDb().Err() return master.FlushDb().Err()
@ -543,11 +478,13 @@ func BenchmarkRedisClusterPing(b *testing.B) {
processes: make(map[string]*redisProcess, 6), processes: make(map[string]*redisProcess, 6),
clients: make(map[string]*redis.Client, 6), clients: make(map[string]*redis.Client, 6),
} }
if err := startCluster(cluster); err != nil { if err := startCluster(cluster); err != nil {
b.Fatal(err) b.Fatal(err)
} }
defer stopCluster(cluster) defer stopCluster(cluster)
client := cluster.clusterClient(nil)
client := cluster.clusterClient(redisClusterOptions())
defer client.Close() defer client.Close()
b.ResetTimer() b.ResetTimer()

View File

@ -1297,7 +1297,7 @@ var _ = Describe("Commands", func() {
stats := client.PoolStats() stats := client.PoolStats()
Expect(stats.Requests).To(Equal(uint32(3))) Expect(stats.Requests).To(Equal(uint32(3)))
Expect(stats.Hits).To(Equal(uint32(2))) Expect(stats.Hits).To(Equal(uint32(1)))
Expect(stats.Timeouts).To(Equal(uint32(0))) Expect(stats.Timeouts).To(Equal(uint32(0)))
}) })

View File

@ -1,10 +1,18 @@
package hashtag package hashtag
import ( import (
"math/rand"
"testing"
. "github.com/onsi/ginkgo" . "github.com/onsi/ginkgo"
. "github.com/onsi/gomega" . "github.com/onsi/gomega"
) )
func TestGinkgoSuite(t *testing.T) {
RegisterFailHandler(Fail)
RunSpecs(t, "hashtag")
}
var _ = Describe("CRC16", func() { var _ = Describe("CRC16", func() {
// http://redis.io/topics/cluster-spec#keys-distribution-model // http://redis.io/topics/cluster-spec#keys-distribution-model
@ -23,3 +31,44 @@ var _ = Describe("CRC16", func() {
}) })
}) })
var _ = Describe("HashSlot", func() {
It("should calculate hash slots", func() {
tests := []struct {
key string
slot int
}{
{"123456789", 12739},
{"{}foo", 9500},
{"foo{}", 5542},
{"foo{}{bar}", 8363},
{"", 10503},
{"", 5176},
{string([]byte{83, 153, 134, 118, 229, 214, 244, 75, 140, 37, 215, 215}), 5463},
}
// Empty keys receive random slot.
rand.Seed(100)
for _, test := range tests {
Expect(Slot(test.key)).To(Equal(test.slot), "for %s", test.key)
}
})
It("should extract keys from tags", func() {
tests := []struct {
one, two string
}{
{"foo{bar}", "bar"},
{"{foo}bar", "foo"},
{"{user1000}.following", "{user1000}.followers"},
{"foo{{bar}}zap", "{bar"},
{"foo{bar}{zap}", "bar"},
}
for _, test := range tests {
Expect(Slot(test.one)).To(Equal(Slot(test.two)), "for %s <-> %s", test.one, test.two)
}
})
})

View File

@ -25,8 +25,8 @@ var timers = sync.Pool{
}, },
} }
// PoolStats contains pool state information and accumulated stats. // Stats contains pool state information and accumulated stats.
type PoolStats struct { type Stats struct {
Requests uint32 // number of times a connection was requested by the pool Requests uint32 // number of times a connection was requested by the pool
Hits uint32 // number of times free connection was found in the pool Hits uint32 // number of times free connection was found in the pool
Timeouts uint32 // number of times a wait timeout occurred Timeouts uint32 // number of times a wait timeout occurred
@ -41,7 +41,7 @@ type Pooler interface {
Remove(*Conn, error) error Remove(*Conn, error) error
Len() int Len() int
FreeLen() int FreeLen() int
Stats() *PoolStats Stats() *Stats
Close() error Close() error
Closed() bool Closed() bool
} }
@ -64,7 +64,7 @@ type ConnPool struct {
freeConnsMu sync.Mutex freeConnsMu sync.Mutex
freeConns []*Conn freeConns []*Conn
stats PoolStats stats Stats
_closed int32 // atomic _closed int32 // atomic
lastErr atomic.Value lastErr atomic.Value
@ -173,16 +173,22 @@ func (p *ConnPool) Get() (*Conn, bool, error) {
return nil, false, ErrPoolTimeout return nil, false, ErrPoolTimeout
} }
p.freeConnsMu.Lock() for {
cn := p.popFree() p.freeConnsMu.Lock()
p.freeConnsMu.Unlock() cn := p.popFree()
p.freeConnsMu.Unlock()
if cn != nil { if cn == nil {
atomic.AddUint32(&p.stats.Hits, 1) break
if !cn.IsStale(p.idleTimeout) {
return cn, false, nil
} }
_ = p.closeConn(cn, errConnStale)
if cn.IsStale(p.idleTimeout) {
p.remove(cn, errConnStale)
continue
}
atomic.AddUint32(&p.stats.Hits, 1)
return cn, false, nil
} }
newcn, err := p.NewConn() newcn, err := p.NewConn()
@ -192,9 +198,6 @@ func (p *ConnPool) Get() (*Conn, bool, error) {
} }
p.connsMu.Lock() p.connsMu.Lock()
if cn != nil {
p.removeConn(cn)
}
p.conns = append(p.conns, newcn) p.conns = append(p.conns, newcn)
p.connsMu.Unlock() p.connsMu.Unlock()
@ -224,17 +227,13 @@ func (p *ConnPool) remove(cn *Conn, reason error) {
_ = p.closeConn(cn, reason) _ = p.closeConn(cn, reason)
p.connsMu.Lock() p.connsMu.Lock()
p.removeConn(cn)
p.connsMu.Unlock()
}
func (p *ConnPool) removeConn(cn *Conn) {
for i, c := range p.conns { for i, c := range p.conns {
if c == cn { if c == cn {
p.conns = append(p.conns[:i], p.conns[i+1:]...) p.conns = append(p.conns[:i], p.conns[i+1:]...)
break break
} }
} }
p.connsMu.Unlock()
} }
// Len returns total number of connections. // Len returns total number of connections.
@ -253,14 +252,14 @@ func (p *ConnPool) FreeLen() int {
return l return l
} }
func (p *ConnPool) Stats() *PoolStats { func (p *ConnPool) Stats() *Stats {
stats := PoolStats{} return &Stats{
stats.Requests = atomic.LoadUint32(&p.stats.Requests) Requests: atomic.LoadUint32(&p.stats.Requests),
stats.Hits = atomic.LoadUint32(&p.stats.Hits) Hits: atomic.LoadUint32(&p.stats.Hits),
stats.Timeouts = atomic.LoadUint32(&p.stats.Timeouts) Timeouts: atomic.LoadUint32(&p.stats.Timeouts),
stats.TotalConns = uint32(p.Len()) TotalConns: uint32(p.Len()),
stats.FreeConns = uint32(p.FreeLen()) FreeConns: uint32(p.FreeLen()),
return &stats }
} }
func (p *ConnPool) Closed() bool { func (p *ConnPool) Closed() bool {

View File

@ -42,7 +42,7 @@ func (p *SingleConnPool) FreeLen() int {
return 0 return 0
} }
func (p *SingleConnPool) Stats() *PoolStats { func (p *SingleConnPool) Stats() *Stats {
return nil return nil
} }

View File

@ -106,7 +106,9 @@ func (p *StickyConnPool) FreeLen() int {
return 0 return 0
} }
func (p *StickyConnPool) Stats() *PoolStats { return nil } func (p *StickyConnPool) Stats() *Stats {
return nil
}
func (p *StickyConnPool) Close() error { func (p *StickyConnPool) Close() error {
defer p.mx.Unlock() defer p.mx.Unlock()

View File

@ -108,8 +108,36 @@ func redisOptions() *redis.Options {
WriteTimeout: 30 * time.Second, WriteTimeout: 30 * time.Second,
PoolSize: 10, PoolSize: 10,
PoolTimeout: 30 * time.Second, PoolTimeout: 30 * time.Second,
IdleTimeout: time.Second, IdleTimeout: 500 * time.Millisecond,
IdleCheckFrequency: time.Second, IdleCheckFrequency: 500 * time.Millisecond,
}
}
func redisClusterOptions() *redis.ClusterOptions {
return &redis.ClusterOptions{
DialTimeout: 10 * time.Second,
ReadTimeout: 30 * time.Second,
WriteTimeout: 30 * time.Second,
PoolSize: 10,
PoolTimeout: 30 * time.Second,
IdleTimeout: 500 * time.Millisecond,
IdleCheckFrequency: 500 * time.Millisecond,
}
}
func redisRingOptions() *redis.RingOptions {
return &redis.RingOptions{
Addrs: map[string]string{
"ringShardOne": ":" + ringShard1Port,
"ringShardTwo": ":" + ringShard2Port,
},
DialTimeout: 10 * time.Second,
ReadTimeout: 30 * time.Second,
WriteTimeout: 30 * time.Second,
PoolSize: 10,
PoolTimeout: 30 * time.Second,
IdleTimeout: 500 * time.Millisecond,
IdleCheckFrequency: 500 * time.Millisecond,
} }
} }

View File

@ -1,6 +1,8 @@
package redis_test package redis_test
import ( import (
"time"
. "github.com/onsi/ginkgo" . "github.com/onsi/ginkgo"
. "github.com/onsi/gomega" . "github.com/onsi/gomega"
@ -20,7 +22,7 @@ var _ = Describe("pool", func() {
Expect(client.Close()).NotTo(HaveOccurred()) Expect(client.Close()).NotTo(HaveOccurred())
}) })
It("should respect max size", func() { It("respects max size", func() {
perform(1000, func(id int) { perform(1000, func(id int) {
val, err := client.Ping().Result() val, err := client.Ping().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
@ -33,7 +35,7 @@ var _ = Describe("pool", func() {
Expect(pool.Len()).To(Equal(pool.FreeLen())) Expect(pool.Len()).To(Equal(pool.FreeLen()))
}) })
It("should respect max on multi", func() { It("srespect max size on multi", func() {
perform(1000, func(id int) { perform(1000, func(id int) {
var ping *redis.StatusCmd var ping *redis.StatusCmd
@ -58,7 +60,7 @@ var _ = Describe("pool", func() {
Expect(pool.Len()).To(Equal(pool.FreeLen())) Expect(pool.Len()).To(Equal(pool.FreeLen()))
}) })
It("should respect max on pipelines", func() { It("respects max size on pipelines", func() {
perform(1000, func(id int) { perform(1000, func(id int) {
pipe := client.Pipeline() pipe := client.Pipeline()
ping := pipe.Ping() ping := pipe.Ping()
@ -76,7 +78,7 @@ var _ = Describe("pool", func() {
Expect(pool.Len()).To(Equal(pool.FreeLen())) Expect(pool.Len()).To(Equal(pool.FreeLen()))
}) })
It("should respect max on pubsub", func() { It("respects max size on pubsub", func() {
connPool := client.Pool() connPool := client.Pool()
connPool.(*pool.ConnPool).DialLimiter = nil connPool.(*pool.ConnPool).DialLimiter = nil
@ -90,7 +92,7 @@ var _ = Describe("pool", func() {
Expect(connPool.Len()).To(BeNumerically("<=", 10)) Expect(connPool.Len()).To(BeNumerically("<=", 10))
}) })
It("should remove broken connections", func() { It("removes broken connections", func() {
cn, _, err := client.Pool().Get() cn, _, err := client.Pool().Get()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
cn.NetConn = &badConn{} cn.NetConn = &badConn{}
@ -113,7 +115,7 @@ var _ = Describe("pool", func() {
Expect(stats.Timeouts).To(Equal(uint32(0))) Expect(stats.Timeouts).To(Equal(uint32(0)))
}) })
It("should reuse connections", func() { It("reuses connections", func() {
for i := 0; i < 100; i++ { for i := 0; i < 100; i++ {
val, err := client.Ping().Result() val, err := client.Ping().Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
@ -129,4 +131,26 @@ var _ = Describe("pool", func() {
Expect(stats.Hits).To(Equal(uint32(100))) Expect(stats.Hits).To(Equal(uint32(100)))
Expect(stats.Timeouts).To(Equal(uint32(0))) Expect(stats.Timeouts).To(Equal(uint32(0)))
}) })
It("removes idle connections", func() {
stats := client.PoolStats()
Expect(stats).To(Equal(&redis.PoolStats{
Requests: 1,
Hits: 0,
Timeouts: 0,
TotalConns: 1,
FreeConns: 1,
}))
time.Sleep(2 * time.Second)
stats = client.PoolStats()
Expect(stats).To(Equal(&redis.PoolStats{
Requests: 1,
Hits: 0,
Timeouts: 0,
TotalConns: 0,
FreeConns: 0,
}))
})
}) })

55
ring.go
View File

@ -4,6 +4,7 @@ import (
"errors" "errors"
"fmt" "fmt"
"sync" "sync"
"sync/atomic"
"time" "time"
"gopkg.in/redis.v4/internal" "gopkg.in/redis.v4/internal"
@ -65,7 +66,7 @@ func (opt *RingOptions) clientOptions() *Options {
type ringShard struct { type ringShard struct {
Client *Client Client *Client
down int down int32
} }
func (shard *ringShard) String() string { func (shard *ringShard) String() string {
@ -80,7 +81,7 @@ func (shard *ringShard) String() string {
func (shard *ringShard) IsDown() bool { func (shard *ringShard) IsDown() bool {
const threshold = 3 const threshold = 3
return shard.down >= threshold return atomic.LoadInt32(&shard.down) >= threshold
} }
func (shard *ringShard) IsUp() bool { func (shard *ringShard) IsUp() bool {
@ -91,7 +92,7 @@ func (shard *ringShard) IsUp() bool {
func (shard *ringShard) Vote(up bool) bool { func (shard *ringShard) Vote(up bool) bool {
if up { if up {
changed := shard.IsDown() changed := shard.IsDown()
shard.down = 0 atomic.StoreInt32(&shard.down, 0)
return changed return changed
} }
@ -99,7 +100,7 @@ func (shard *ringShard) Vote(up bool) bool {
return false return false
} }
shard.down++ atomic.AddInt32(&shard.down, 1)
return shard.IsDown() return shard.IsDown()
} }
@ -157,6 +158,52 @@ func NewRing(opt *RingOptions) *Ring {
return ring return ring
} }
// PoolStats returns accumulated connection pool stats.
func (c *Ring) PoolStats() *PoolStats {
var acc PoolStats
for _, shard := range c.shards {
s := shard.Client.connPool.Stats()
acc.Requests += s.Requests
acc.Hits += s.Hits
acc.Timeouts += s.Timeouts
acc.TotalConns += s.TotalConns
acc.FreeConns += s.FreeConns
}
return &acc
}
// ForEachShard concurrently calls the fn on each live shard in the ring.
// It returns the first error if any.
func (c *Ring) ForEachShard(fn func(client *Client) error) error {
var wg sync.WaitGroup
errCh := make(chan error, 1)
for _, shard := range c.shards {
if shard.IsDown() {
continue
}
wg.Add(1)
go func(shard *ringShard) {
defer wg.Done()
err := fn(shard.Client)
if err != nil {
select {
case errCh <- err:
default:
}
}
}(shard)
}
wg.Wait()
select {
case err := <-errCh:
return err
default:
return nil
}
}
func (c *Ring) cmdInfo(name string) *CommandInfo { func (c *Ring) cmdInfo(name string) *CommandInfo {
c.cmdsInfoOnce.Do(func() { c.cmdsInfoOnce.Do(func() {
for _, shard := range c.shards { for _, shard := range c.shards {

View File

@ -11,8 +11,9 @@ import (
"gopkg.in/redis.v4" "gopkg.in/redis.v4"
) )
var _ = Describe("Redis ring", func() { var _ = Describe("Redis Ring", func() {
const heartbeat = 100 * time.Millisecond const heartbeat = 100 * time.Millisecond
var ring *redis.Ring var ring *redis.Ring
setRingKeys := func() { setRingKeys := func() {
@ -23,20 +24,14 @@ var _ = Describe("Redis ring", func() {
} }
BeforeEach(func() { BeforeEach(func() {
ring = redis.NewRing(&redis.RingOptions{ opt := redisRingOptions()
Addrs: map[string]string{ opt.HeartbeatFrequency = heartbeat
"ringShardOne": ":" + ringShard1Port, ring = redis.NewRing(opt)
"ringShardTwo": ":" + ringShard2Port,
}, err := ring.ForEachShard(func(cl *redis.Client) error {
HeartbeatFrequency: heartbeat, return cl.FlushDb().Err()
}) })
Expect(err).NotTo(HaveOccurred())
// Shards should not have any keys.
Expect(ringShard1.FlushDb().Err()).NotTo(HaveOccurred())
Expect(ringShard1.Info().Val()).NotTo(ContainSubstring("keys="))
Expect(ringShard2.FlushDb().Err()).NotTo(HaveOccurred())
Expect(ringShard2.Info().Val()).NotTo(ContainSubstring("keys="))
}) })
AfterEach(func() { AfterEach(func() {