forked from mirror/redis
Enable reaper on ClusterClient and add tests.
This commit is contained in:
parent
a7d1d0b9ac
commit
7cbee9d337
39
cluster.go
39
cluster.go
|
@ -45,20 +45,25 @@ var _ Cmdable = (*ClusterClient)(nil)
|
||||||
// http://redis.io/topics/cluster-spec.
|
// http://redis.io/topics/cluster-spec.
|
||||||
func NewClusterClient(opt *ClusterOptions) *ClusterClient {
|
func NewClusterClient(opt *ClusterOptions) *ClusterClient {
|
||||||
opt.init()
|
opt.init()
|
||||||
client := &ClusterClient{
|
|
||||||
|
c := &ClusterClient{
|
||||||
opt: opt,
|
opt: opt,
|
||||||
nodes: make(map[string]*clusterNode),
|
nodes: make(map[string]*clusterNode),
|
||||||
|
|
||||||
cmdsInfoOnce: new(sync.Once),
|
cmdsInfoOnce: new(sync.Once),
|
||||||
}
|
}
|
||||||
client.cmdable.process = client.Process
|
c.cmdable.process = c.Process
|
||||||
|
|
||||||
for _, addr := range opt.Addrs {
|
for _, addr := range opt.Addrs {
|
||||||
_, _ = client.nodeByAddr(addr)
|
_, _ = c.nodeByAddr(addr)
|
||||||
}
|
}
|
||||||
client.reloadSlots()
|
c.reloadSlots()
|
||||||
|
|
||||||
return client
|
if opt.IdleCheckFrequency > 0 {
|
||||||
|
go c.reaper(opt.IdleCheckFrequency)
|
||||||
|
}
|
||||||
|
|
||||||
|
return c
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *ClusterClient) cmdInfo(name string) *CommandInfo {
|
func (c *ClusterClient) cmdInfo(name string) *CommandInfo {
|
||||||
|
@ -333,11 +338,9 @@ func (c *ClusterClient) ForEachMaster(fn func(client *Client) error) error {
|
||||||
slots := c.slots
|
slots := c.slots
|
||||||
c.mu.RUnlock()
|
c.mu.RUnlock()
|
||||||
|
|
||||||
var retErr error
|
|
||||||
var mu sync.Mutex
|
|
||||||
|
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
visited := make(map[*clusterNode]struct{})
|
visited := make(map[*clusterNode]struct{})
|
||||||
|
errCh := make(chan error, 1)
|
||||||
for _, nodes := range slots {
|
for _, nodes := range slots {
|
||||||
if len(nodes) == 0 {
|
if len(nodes) == 0 {
|
||||||
continue
|
continue
|
||||||
|
@ -351,20 +354,24 @@ func (c *ClusterClient) ForEachMaster(fn func(client *Client) error) error {
|
||||||
|
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func(node *clusterNode) {
|
go func(node *clusterNode) {
|
||||||
|
defer wg.Done()
|
||||||
err := fn(node.Client)
|
err := fn(node.Client)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
mu.Lock()
|
select {
|
||||||
if retErr == nil {
|
case errCh <- err:
|
||||||
retErr = err
|
default:
|
||||||
}
|
}
|
||||||
mu.Unlock()
|
|
||||||
}
|
}
|
||||||
wg.Done()
|
|
||||||
}(master)
|
}(master)
|
||||||
}
|
}
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
||||||
return retErr
|
select {
|
||||||
|
case err := <-errCh:
|
||||||
|
return err
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// closeClients closes all clients and returns the first error if there are any.
|
// closeClients closes all clients and returns the first error if there are any.
|
||||||
|
@ -442,8 +449,8 @@ func (c *ClusterClient) setNodesLatency() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// reaper closes idle connections to the cluster.
|
// reaper closes idle connections to the cluster.
|
||||||
func (c *ClusterClient) reaper(frequency time.Duration) {
|
func (c *ClusterClient) reaper(idleCheckFrequency time.Duration) {
|
||||||
ticker := time.NewTicker(frequency)
|
ticker := time.NewTicker(idleCheckFrequency)
|
||||||
defer ticker.Stop()
|
defer ticker.Stop()
|
||||||
|
|
||||||
for _ = range ticker.C {
|
for _ = range ticker.C {
|
||||||
|
|
129
cluster_test.go
129
cluster_test.go
|
@ -2,7 +2,6 @@ package redis_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/rand"
|
|
||||||
"net"
|
"net"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
@ -50,17 +49,6 @@ func (s *clusterScenario) clusterClient(opt *redis.ClusterOptions) *redis.Cluste
|
||||||
for i, port := range s.ports {
|
for i, port := range s.ports {
|
||||||
addrs[i] = net.JoinHostPort("127.0.0.1", port)
|
addrs[i] = net.JoinHostPort("127.0.0.1", port)
|
||||||
}
|
}
|
||||||
if opt == nil {
|
|
||||||
opt = &redis.ClusterOptions{
|
|
||||||
DialTimeout: 10 * time.Second,
|
|
||||||
ReadTimeout: 30 * time.Second,
|
|
||||||
WriteTimeout: 30 * time.Second,
|
|
||||||
PoolSize: 10,
|
|
||||||
PoolTimeout: 30 * time.Second,
|
|
||||||
IdleTimeout: time.Second,
|
|
||||||
IdleCheckFrequency: time.Second,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
opt.Addrs = addrs
|
opt.Addrs = addrs
|
||||||
return redis.NewClusterClient(opt)
|
return redis.NewClusterClient(opt)
|
||||||
}
|
}
|
||||||
|
@ -193,52 +181,12 @@ func stopCluster(scenario *clusterScenario) error {
|
||||||
|
|
||||||
//------------------------------------------------------------------------------
|
//------------------------------------------------------------------------------
|
||||||
|
|
||||||
var _ = Describe("Cluster", func() {
|
var _ = Describe("ClusterClient", func() {
|
||||||
Describe("HashSlot", func() {
|
var client *redis.ClusterClient
|
||||||
|
|
||||||
It("should calculate hash slots", func() {
|
|
||||||
tests := []struct {
|
|
||||||
key string
|
|
||||||
slot int
|
|
||||||
}{
|
|
||||||
{"123456789", 12739},
|
|
||||||
{"{}foo", 9500},
|
|
||||||
{"foo{}", 5542},
|
|
||||||
{"foo{}{bar}", 8363},
|
|
||||||
{"", 10503},
|
|
||||||
{"", 5176},
|
|
||||||
{string([]byte{83, 153, 134, 118, 229, 214, 244, 75, 140, 37, 215, 215}), 5463},
|
|
||||||
}
|
|
||||||
// Empty keys receive random slot.
|
|
||||||
rand.Seed(100)
|
|
||||||
|
|
||||||
for _, test := range tests {
|
|
||||||
Expect(hashtag.Slot(test.key)).To(Equal(test.slot), "for %s", test.key)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
It("should extract keys from tags", func() {
|
|
||||||
tests := []struct {
|
|
||||||
one, two string
|
|
||||||
}{
|
|
||||||
{"foo{bar}", "bar"},
|
|
||||||
{"{foo}bar", "foo"},
|
|
||||||
{"{user1000}.following", "{user1000}.followers"},
|
|
||||||
{"foo{{bar}}zap", "{bar"},
|
|
||||||
{"foo{bar}{zap}", "bar"},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, test := range tests {
|
|
||||||
Expect(hashtag.Slot(test.one)).To(Equal(hashtag.Slot(test.two)), "for %s <-> %s", test.one, test.two)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
})
|
|
||||||
|
|
||||||
Describe("Commands", func() {
|
|
||||||
|
|
||||||
|
describeClusterClient := func() {
|
||||||
It("should CLUSTER SLOTS", func() {
|
It("should CLUSTER SLOTS", func() {
|
||||||
res, err := cluster.primary().ClusterSlots().Result()
|
res, err := client.ClusterSlots().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(res).To(HaveLen(3))
|
Expect(res).To(HaveLen(3))
|
||||||
|
|
||||||
|
@ -251,73 +199,48 @@ var _ = Describe("Cluster", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER NODES", func() {
|
It("should CLUSTER NODES", func() {
|
||||||
res, err := cluster.primary().ClusterNodes().Result()
|
res, err := client.ClusterNodes().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(len(res)).To(BeNumerically(">", 400))
|
Expect(len(res)).To(BeNumerically(">", 400))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER INFO", func() {
|
It("should CLUSTER INFO", func() {
|
||||||
res, err := cluster.primary().ClusterInfo().Result()
|
res, err := client.ClusterInfo().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(res).To(ContainSubstring("cluster_known_nodes:6"))
|
Expect(res).To(ContainSubstring("cluster_known_nodes:6"))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER KEYSLOT", func() {
|
It("should CLUSTER KEYSLOT", func() {
|
||||||
hashSlot, err := cluster.primary().ClusterKeySlot("somekey").Result()
|
hashSlot, err := client.ClusterKeySlot("somekey").Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey"))))
|
Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey"))))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER COUNT-FAILURE-REPORTS", func() {
|
It("should CLUSTER COUNT-FAILURE-REPORTS", func() {
|
||||||
n, err := cluster.primary().ClusterCountFailureReports(cluster.nodeIds[0]).Result()
|
n, err := client.ClusterCountFailureReports(cluster.nodeIds[0]).Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(n).To(Equal(int64(0)))
|
Expect(n).To(Equal(int64(0)))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER COUNTKEYSINSLOT", func() {
|
It("should CLUSTER COUNTKEYSINSLOT", func() {
|
||||||
n, err := cluster.primary().ClusterCountKeysInSlot(10).Result()
|
n, err := client.ClusterCountKeysInSlot(10).Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(n).To(Equal(int64(0)))
|
Expect(n).To(Equal(int64(0)))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER DELSLOTS", func() {
|
|
||||||
res, err := cluster.primary().ClusterDelSlotsRange(16000, 16384-1).Result()
|
|
||||||
Expect(err).NotTo(HaveOccurred())
|
|
||||||
Expect(res).To(Equal("OK"))
|
|
||||||
cluster.primary().ClusterAddSlotsRange(16000, 16384-1)
|
|
||||||
})
|
|
||||||
|
|
||||||
It("should CLUSTER SAVECONFIG", func() {
|
It("should CLUSTER SAVECONFIG", func() {
|
||||||
res, err := cluster.primary().ClusterSaveConfig().Result()
|
res, err := client.ClusterSaveConfig().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(res).To(Equal("OK"))
|
Expect(res).To(Equal("OK"))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should CLUSTER SLAVES", func() {
|
It("should CLUSTER SLAVES", func() {
|
||||||
nodesList, err := cluster.primary().ClusterSlaves(cluster.nodeIds[0]).Result()
|
nodesList, err := client.ClusterSlaves(cluster.nodeIds[0]).Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
Expect(nodesList).Should(ContainElement(ContainSubstring("slave")))
|
Expect(nodesList).Should(ContainElement(ContainSubstring("slave")))
|
||||||
Expect(nodesList).Should(HaveLen(1))
|
Expect(nodesList).Should(HaveLen(1))
|
||||||
})
|
})
|
||||||
|
|
||||||
// It("should CLUSTER READONLY", func() {
|
|
||||||
// res, err := cluster.primary().ReadOnly().Result()
|
|
||||||
// Expect(err).NotTo(HaveOccurred())
|
|
||||||
// Expect(res).To(Equal("OK"))
|
|
||||||
// })
|
|
||||||
|
|
||||||
// It("should CLUSTER READWRITE", func() {
|
|
||||||
// res, err := cluster.primary().ReadWrite().Result()
|
|
||||||
// Expect(err).NotTo(HaveOccurred())
|
|
||||||
// Expect(res).To(Equal("OK"))
|
|
||||||
// })
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
var _ = Describe("ClusterClient", func() {
|
|
||||||
var client *redis.ClusterClient
|
|
||||||
|
|
||||||
describeClusterClient := func() {
|
|
||||||
It("should GET/SET/DEL", func() {
|
It("should GET/SET/DEL", func() {
|
||||||
val, err := client.Get("A").Result()
|
val, err := client.Get("A").Result()
|
||||||
Expect(err).To(Equal(redis.Nil))
|
Expect(err).To(Equal(redis.Nil))
|
||||||
|
@ -340,6 +263,18 @@ var _ = Describe("ClusterClient", func() {
|
||||||
Expect(client.PoolStats()).To(BeAssignableToTypeOf(&redis.PoolStats{}))
|
Expect(client.PoolStats()).To(BeAssignableToTypeOf(&redis.PoolStats{}))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
It("removes idle connections", func() {
|
||||||
|
stats := client.PoolStats()
|
||||||
|
Expect(stats.TotalConns).NotTo(BeZero())
|
||||||
|
Expect(stats.FreeConns).NotTo(BeZero())
|
||||||
|
|
||||||
|
time.Sleep(2 * time.Second)
|
||||||
|
|
||||||
|
stats = client.PoolStats()
|
||||||
|
Expect(stats.TotalConns).To(BeZero())
|
||||||
|
Expect(stats.FreeConns).To(BeZero())
|
||||||
|
})
|
||||||
|
|
||||||
It("follows redirects", func() {
|
It("follows redirects", func() {
|
||||||
Expect(client.Set("A", "VALUE", 0).Err()).NotTo(HaveOccurred())
|
Expect(client.Set("A", "VALUE", 0).Err()).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
@ -352,9 +287,9 @@ var _ = Describe("ClusterClient", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
It("returns an error when there are no attempts left", func() {
|
It("returns an error when there are no attempts left", func() {
|
||||||
client := cluster.clusterClient(&redis.ClusterOptions{
|
opt := redisClusterOptions()
|
||||||
MaxRedirects: -1,
|
opt.MaxRedirects = -1
|
||||||
})
|
client := cluster.clusterClient(opt)
|
||||||
|
|
||||||
slot := hashtag.Slot("A")
|
slot := hashtag.Slot("A")
|
||||||
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
|
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
|
||||||
|
@ -483,7 +418,7 @@ var _ = Describe("ClusterClient", func() {
|
||||||
|
|
||||||
Describe("default ClusterClient", func() {
|
Describe("default ClusterClient", func() {
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
client = cluster.clusterClient(nil)
|
client = cluster.clusterClient(redisClusterOptions())
|
||||||
|
|
||||||
_ = client.ForEachMaster(func(master *redis.Client) error {
|
_ = client.ForEachMaster(func(master *redis.Client) error {
|
||||||
return master.FlushDb().Err()
|
return master.FlushDb().Err()
|
||||||
|
@ -499,9 +434,9 @@ var _ = Describe("ClusterClient", func() {
|
||||||
|
|
||||||
Describe("ClusterClient with RouteByLatency", func() {
|
Describe("ClusterClient with RouteByLatency", func() {
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
client = cluster.clusterClient(&redis.ClusterOptions{
|
opt := redisClusterOptions()
|
||||||
RouteByLatency: true,
|
opt.RouteByLatency = true
|
||||||
})
|
client = cluster.clusterClient(opt)
|
||||||
|
|
||||||
_ = client.ForEachMaster(func(master *redis.Client) error {
|
_ = client.ForEachMaster(func(master *redis.Client) error {
|
||||||
return master.FlushDb().Err()
|
return master.FlushDb().Err()
|
||||||
|
@ -543,11 +478,13 @@ func BenchmarkRedisClusterPing(b *testing.B) {
|
||||||
processes: make(map[string]*redisProcess, 6),
|
processes: make(map[string]*redisProcess, 6),
|
||||||
clients: make(map[string]*redis.Client, 6),
|
clients: make(map[string]*redis.Client, 6),
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := startCluster(cluster); err != nil {
|
if err := startCluster(cluster); err != nil {
|
||||||
b.Fatal(err)
|
b.Fatal(err)
|
||||||
}
|
}
|
||||||
defer stopCluster(cluster)
|
defer stopCluster(cluster)
|
||||||
client := cluster.clusterClient(nil)
|
|
||||||
|
client := cluster.clusterClient(redisClusterOptions())
|
||||||
defer client.Close()
|
defer client.Close()
|
||||||
|
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
|
|
|
@ -1297,7 +1297,7 @@ var _ = Describe("Commands", func() {
|
||||||
|
|
||||||
stats := client.PoolStats()
|
stats := client.PoolStats()
|
||||||
Expect(stats.Requests).To(Equal(uint32(3)))
|
Expect(stats.Requests).To(Equal(uint32(3)))
|
||||||
Expect(stats.Hits).To(Equal(uint32(2)))
|
Expect(stats.Hits).To(Equal(uint32(1)))
|
||||||
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,18 @@
|
||||||
package hashtag
|
package hashtag
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"math/rand"
|
||||||
|
"testing"
|
||||||
|
|
||||||
. "github.com/onsi/ginkgo"
|
. "github.com/onsi/ginkgo"
|
||||||
. "github.com/onsi/gomega"
|
. "github.com/onsi/gomega"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
func TestGinkgoSuite(t *testing.T) {
|
||||||
|
RegisterFailHandler(Fail)
|
||||||
|
RunSpecs(t, "hashtag")
|
||||||
|
}
|
||||||
|
|
||||||
var _ = Describe("CRC16", func() {
|
var _ = Describe("CRC16", func() {
|
||||||
|
|
||||||
// http://redis.io/topics/cluster-spec#keys-distribution-model
|
// http://redis.io/topics/cluster-spec#keys-distribution-model
|
||||||
|
@ -23,3 +31,44 @@ var _ = Describe("CRC16", func() {
|
||||||
})
|
})
|
||||||
|
|
||||||
})
|
})
|
||||||
|
|
||||||
|
var _ = Describe("HashSlot", func() {
|
||||||
|
|
||||||
|
It("should calculate hash slots", func() {
|
||||||
|
tests := []struct {
|
||||||
|
key string
|
||||||
|
slot int
|
||||||
|
}{
|
||||||
|
{"123456789", 12739},
|
||||||
|
{"{}foo", 9500},
|
||||||
|
{"foo{}", 5542},
|
||||||
|
{"foo{}{bar}", 8363},
|
||||||
|
{"", 10503},
|
||||||
|
{"", 5176},
|
||||||
|
{string([]byte{83, 153, 134, 118, 229, 214, 244, 75, 140, 37, 215, 215}), 5463},
|
||||||
|
}
|
||||||
|
// Empty keys receive random slot.
|
||||||
|
rand.Seed(100)
|
||||||
|
|
||||||
|
for _, test := range tests {
|
||||||
|
Expect(Slot(test.key)).To(Equal(test.slot), "for %s", test.key)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
It("should extract keys from tags", func() {
|
||||||
|
tests := []struct {
|
||||||
|
one, two string
|
||||||
|
}{
|
||||||
|
{"foo{bar}", "bar"},
|
||||||
|
{"{foo}bar", "foo"},
|
||||||
|
{"{user1000}.following", "{user1000}.followers"},
|
||||||
|
{"foo{{bar}}zap", "{bar"},
|
||||||
|
{"foo{bar}{zap}", "bar"},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range tests {
|
||||||
|
Expect(Slot(test.one)).To(Equal(Slot(test.two)), "for %s <-> %s", test.one, test.two)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
})
|
||||||
|
|
|
@ -25,8 +25,8 @@ var timers = sync.Pool{
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
// PoolStats contains pool state information and accumulated stats.
|
// Stats contains pool state information and accumulated stats.
|
||||||
type PoolStats struct {
|
type Stats struct {
|
||||||
Requests uint32 // number of times a connection was requested by the pool
|
Requests uint32 // number of times a connection was requested by the pool
|
||||||
Hits uint32 // number of times free connection was found in the pool
|
Hits uint32 // number of times free connection was found in the pool
|
||||||
Timeouts uint32 // number of times a wait timeout occurred
|
Timeouts uint32 // number of times a wait timeout occurred
|
||||||
|
@ -41,7 +41,7 @@ type Pooler interface {
|
||||||
Remove(*Conn, error) error
|
Remove(*Conn, error) error
|
||||||
Len() int
|
Len() int
|
||||||
FreeLen() int
|
FreeLen() int
|
||||||
Stats() *PoolStats
|
Stats() *Stats
|
||||||
Close() error
|
Close() error
|
||||||
Closed() bool
|
Closed() bool
|
||||||
}
|
}
|
||||||
|
@ -64,7 +64,7 @@ type ConnPool struct {
|
||||||
freeConnsMu sync.Mutex
|
freeConnsMu sync.Mutex
|
||||||
freeConns []*Conn
|
freeConns []*Conn
|
||||||
|
|
||||||
stats PoolStats
|
stats Stats
|
||||||
|
|
||||||
_closed int32 // atomic
|
_closed int32 // atomic
|
||||||
lastErr atomic.Value
|
lastErr atomic.Value
|
||||||
|
@ -173,16 +173,22 @@ func (p *ConnPool) Get() (*Conn, bool, error) {
|
||||||
return nil, false, ErrPoolTimeout
|
return nil, false, ErrPoolTimeout
|
||||||
}
|
}
|
||||||
|
|
||||||
p.freeConnsMu.Lock()
|
for {
|
||||||
cn := p.popFree()
|
p.freeConnsMu.Lock()
|
||||||
p.freeConnsMu.Unlock()
|
cn := p.popFree()
|
||||||
|
p.freeConnsMu.Unlock()
|
||||||
|
|
||||||
if cn != nil {
|
if cn == nil {
|
||||||
atomic.AddUint32(&p.stats.Hits, 1)
|
break
|
||||||
if !cn.IsStale(p.idleTimeout) {
|
|
||||||
return cn, false, nil
|
|
||||||
}
|
}
|
||||||
_ = p.closeConn(cn, errConnStale)
|
|
||||||
|
if cn.IsStale(p.idleTimeout) {
|
||||||
|
p.remove(cn, errConnStale)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
atomic.AddUint32(&p.stats.Hits, 1)
|
||||||
|
return cn, false, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
newcn, err := p.NewConn()
|
newcn, err := p.NewConn()
|
||||||
|
@ -192,9 +198,6 @@ func (p *ConnPool) Get() (*Conn, bool, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
p.connsMu.Lock()
|
p.connsMu.Lock()
|
||||||
if cn != nil {
|
|
||||||
p.removeConn(cn)
|
|
||||||
}
|
|
||||||
p.conns = append(p.conns, newcn)
|
p.conns = append(p.conns, newcn)
|
||||||
p.connsMu.Unlock()
|
p.connsMu.Unlock()
|
||||||
|
|
||||||
|
@ -224,17 +227,13 @@ func (p *ConnPool) remove(cn *Conn, reason error) {
|
||||||
_ = p.closeConn(cn, reason)
|
_ = p.closeConn(cn, reason)
|
||||||
|
|
||||||
p.connsMu.Lock()
|
p.connsMu.Lock()
|
||||||
p.removeConn(cn)
|
|
||||||
p.connsMu.Unlock()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (p *ConnPool) removeConn(cn *Conn) {
|
|
||||||
for i, c := range p.conns {
|
for i, c := range p.conns {
|
||||||
if c == cn {
|
if c == cn {
|
||||||
p.conns = append(p.conns[:i], p.conns[i+1:]...)
|
p.conns = append(p.conns[:i], p.conns[i+1:]...)
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
p.connsMu.Unlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Len returns total number of connections.
|
// Len returns total number of connections.
|
||||||
|
@ -253,14 +252,14 @@ func (p *ConnPool) FreeLen() int {
|
||||||
return l
|
return l
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *ConnPool) Stats() *PoolStats {
|
func (p *ConnPool) Stats() *Stats {
|
||||||
stats := PoolStats{}
|
return &Stats{
|
||||||
stats.Requests = atomic.LoadUint32(&p.stats.Requests)
|
Requests: atomic.LoadUint32(&p.stats.Requests),
|
||||||
stats.Hits = atomic.LoadUint32(&p.stats.Hits)
|
Hits: atomic.LoadUint32(&p.stats.Hits),
|
||||||
stats.Timeouts = atomic.LoadUint32(&p.stats.Timeouts)
|
Timeouts: atomic.LoadUint32(&p.stats.Timeouts),
|
||||||
stats.TotalConns = uint32(p.Len())
|
TotalConns: uint32(p.Len()),
|
||||||
stats.FreeConns = uint32(p.FreeLen())
|
FreeConns: uint32(p.FreeLen()),
|
||||||
return &stats
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *ConnPool) Closed() bool {
|
func (p *ConnPool) Closed() bool {
|
||||||
|
|
|
@ -42,7 +42,7 @@ func (p *SingleConnPool) FreeLen() int {
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *SingleConnPool) Stats() *PoolStats {
|
func (p *SingleConnPool) Stats() *Stats {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -106,7 +106,9 @@ func (p *StickyConnPool) FreeLen() int {
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *StickyConnPool) Stats() *PoolStats { return nil }
|
func (p *StickyConnPool) Stats() *Stats {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (p *StickyConnPool) Close() error {
|
func (p *StickyConnPool) Close() error {
|
||||||
defer p.mx.Unlock()
|
defer p.mx.Unlock()
|
||||||
|
|
32
main_test.go
32
main_test.go
|
@ -108,8 +108,36 @@ func redisOptions() *redis.Options {
|
||||||
WriteTimeout: 30 * time.Second,
|
WriteTimeout: 30 * time.Second,
|
||||||
PoolSize: 10,
|
PoolSize: 10,
|
||||||
PoolTimeout: 30 * time.Second,
|
PoolTimeout: 30 * time.Second,
|
||||||
IdleTimeout: time.Second,
|
IdleTimeout: 500 * time.Millisecond,
|
||||||
IdleCheckFrequency: time.Second,
|
IdleCheckFrequency: 500 * time.Millisecond,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func redisClusterOptions() *redis.ClusterOptions {
|
||||||
|
return &redis.ClusterOptions{
|
||||||
|
DialTimeout: 10 * time.Second,
|
||||||
|
ReadTimeout: 30 * time.Second,
|
||||||
|
WriteTimeout: 30 * time.Second,
|
||||||
|
PoolSize: 10,
|
||||||
|
PoolTimeout: 30 * time.Second,
|
||||||
|
IdleTimeout: 500 * time.Millisecond,
|
||||||
|
IdleCheckFrequency: 500 * time.Millisecond,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func redisRingOptions() *redis.RingOptions {
|
||||||
|
return &redis.RingOptions{
|
||||||
|
Addrs: map[string]string{
|
||||||
|
"ringShardOne": ":" + ringShard1Port,
|
||||||
|
"ringShardTwo": ":" + ringShard2Port,
|
||||||
|
},
|
||||||
|
DialTimeout: 10 * time.Second,
|
||||||
|
ReadTimeout: 30 * time.Second,
|
||||||
|
WriteTimeout: 30 * time.Second,
|
||||||
|
PoolSize: 10,
|
||||||
|
PoolTimeout: 30 * time.Second,
|
||||||
|
IdleTimeout: 500 * time.Millisecond,
|
||||||
|
IdleCheckFrequency: 500 * time.Millisecond,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
36
pool_test.go
36
pool_test.go
|
@ -1,6 +1,8 @@
|
||||||
package redis_test
|
package redis_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
. "github.com/onsi/ginkgo"
|
. "github.com/onsi/ginkgo"
|
||||||
. "github.com/onsi/gomega"
|
. "github.com/onsi/gomega"
|
||||||
|
|
||||||
|
@ -20,7 +22,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(client.Close()).NotTo(HaveOccurred())
|
Expect(client.Close()).NotTo(HaveOccurred())
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should respect max size", func() {
|
It("respects max size", func() {
|
||||||
perform(1000, func(id int) {
|
perform(1000, func(id int) {
|
||||||
val, err := client.Ping().Result()
|
val, err := client.Ping().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
@ -33,7 +35,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should respect max on multi", func() {
|
It("srespect max size on multi", func() {
|
||||||
perform(1000, func(id int) {
|
perform(1000, func(id int) {
|
||||||
var ping *redis.StatusCmd
|
var ping *redis.StatusCmd
|
||||||
|
|
||||||
|
@ -58,7 +60,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should respect max on pipelines", func() {
|
It("respects max size on pipelines", func() {
|
||||||
perform(1000, func(id int) {
|
perform(1000, func(id int) {
|
||||||
pipe := client.Pipeline()
|
pipe := client.Pipeline()
|
||||||
ping := pipe.Ping()
|
ping := pipe.Ping()
|
||||||
|
@ -76,7 +78,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
Expect(pool.Len()).To(Equal(pool.FreeLen()))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should respect max on pubsub", func() {
|
It("respects max size on pubsub", func() {
|
||||||
connPool := client.Pool()
|
connPool := client.Pool()
|
||||||
connPool.(*pool.ConnPool).DialLimiter = nil
|
connPool.(*pool.ConnPool).DialLimiter = nil
|
||||||
|
|
||||||
|
@ -90,7 +92,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(connPool.Len()).To(BeNumerically("<=", 10))
|
Expect(connPool.Len()).To(BeNumerically("<=", 10))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should remove broken connections", func() {
|
It("removes broken connections", func() {
|
||||||
cn, _, err := client.Pool().Get()
|
cn, _, err := client.Pool().Get()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
cn.NetConn = &badConn{}
|
cn.NetConn = &badConn{}
|
||||||
|
@ -113,7 +115,7 @@ var _ = Describe("pool", func() {
|
||||||
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
||||||
})
|
})
|
||||||
|
|
||||||
It("should reuse connections", func() {
|
It("reuses connections", func() {
|
||||||
for i := 0; i < 100; i++ {
|
for i := 0; i < 100; i++ {
|
||||||
val, err := client.Ping().Result()
|
val, err := client.Ping().Result()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
@ -129,4 +131,26 @@ var _ = Describe("pool", func() {
|
||||||
Expect(stats.Hits).To(Equal(uint32(100)))
|
Expect(stats.Hits).To(Equal(uint32(100)))
|
||||||
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
Expect(stats.Timeouts).To(Equal(uint32(0)))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
It("removes idle connections", func() {
|
||||||
|
stats := client.PoolStats()
|
||||||
|
Expect(stats).To(Equal(&redis.PoolStats{
|
||||||
|
Requests: 1,
|
||||||
|
Hits: 0,
|
||||||
|
Timeouts: 0,
|
||||||
|
TotalConns: 1,
|
||||||
|
FreeConns: 1,
|
||||||
|
}))
|
||||||
|
|
||||||
|
time.Sleep(2 * time.Second)
|
||||||
|
|
||||||
|
stats = client.PoolStats()
|
||||||
|
Expect(stats).To(Equal(&redis.PoolStats{
|
||||||
|
Requests: 1,
|
||||||
|
Hits: 0,
|
||||||
|
Timeouts: 0,
|
||||||
|
TotalConns: 0,
|
||||||
|
FreeConns: 0,
|
||||||
|
}))
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|
55
ring.go
55
ring.go
|
@ -4,6 +4,7 @@ import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"sync"
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"gopkg.in/redis.v4/internal"
|
"gopkg.in/redis.v4/internal"
|
||||||
|
@ -65,7 +66,7 @@ func (opt *RingOptions) clientOptions() *Options {
|
||||||
|
|
||||||
type ringShard struct {
|
type ringShard struct {
|
||||||
Client *Client
|
Client *Client
|
||||||
down int
|
down int32
|
||||||
}
|
}
|
||||||
|
|
||||||
func (shard *ringShard) String() string {
|
func (shard *ringShard) String() string {
|
||||||
|
@ -80,7 +81,7 @@ func (shard *ringShard) String() string {
|
||||||
|
|
||||||
func (shard *ringShard) IsDown() bool {
|
func (shard *ringShard) IsDown() bool {
|
||||||
const threshold = 3
|
const threshold = 3
|
||||||
return shard.down >= threshold
|
return atomic.LoadInt32(&shard.down) >= threshold
|
||||||
}
|
}
|
||||||
|
|
||||||
func (shard *ringShard) IsUp() bool {
|
func (shard *ringShard) IsUp() bool {
|
||||||
|
@ -91,7 +92,7 @@ func (shard *ringShard) IsUp() bool {
|
||||||
func (shard *ringShard) Vote(up bool) bool {
|
func (shard *ringShard) Vote(up bool) bool {
|
||||||
if up {
|
if up {
|
||||||
changed := shard.IsDown()
|
changed := shard.IsDown()
|
||||||
shard.down = 0
|
atomic.StoreInt32(&shard.down, 0)
|
||||||
return changed
|
return changed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -99,7 +100,7 @@ func (shard *ringShard) Vote(up bool) bool {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
shard.down++
|
atomic.AddInt32(&shard.down, 1)
|
||||||
return shard.IsDown()
|
return shard.IsDown()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -157,6 +158,52 @@ func NewRing(opt *RingOptions) *Ring {
|
||||||
return ring
|
return ring
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// PoolStats returns accumulated connection pool stats.
|
||||||
|
func (c *Ring) PoolStats() *PoolStats {
|
||||||
|
var acc PoolStats
|
||||||
|
for _, shard := range c.shards {
|
||||||
|
s := shard.Client.connPool.Stats()
|
||||||
|
acc.Requests += s.Requests
|
||||||
|
acc.Hits += s.Hits
|
||||||
|
acc.Timeouts += s.Timeouts
|
||||||
|
acc.TotalConns += s.TotalConns
|
||||||
|
acc.FreeConns += s.FreeConns
|
||||||
|
}
|
||||||
|
return &acc
|
||||||
|
}
|
||||||
|
|
||||||
|
// ForEachShard concurrently calls the fn on each live shard in the ring.
|
||||||
|
// It returns the first error if any.
|
||||||
|
func (c *Ring) ForEachShard(fn func(client *Client) error) error {
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
errCh := make(chan error, 1)
|
||||||
|
for _, shard := range c.shards {
|
||||||
|
if shard.IsDown() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
wg.Add(1)
|
||||||
|
go func(shard *ringShard) {
|
||||||
|
defer wg.Done()
|
||||||
|
err := fn(shard.Client)
|
||||||
|
if err != nil {
|
||||||
|
select {
|
||||||
|
case errCh <- err:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}(shard)
|
||||||
|
}
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
|
select {
|
||||||
|
case err := <-errCh:
|
||||||
|
return err
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Ring) cmdInfo(name string) *CommandInfo {
|
func (c *Ring) cmdInfo(name string) *CommandInfo {
|
||||||
c.cmdsInfoOnce.Do(func() {
|
c.cmdsInfoOnce.Do(func() {
|
||||||
for _, shard := range c.shards {
|
for _, shard := range c.shards {
|
||||||
|
|
23
ring_test.go
23
ring_test.go
|
@ -11,8 +11,9 @@ import (
|
||||||
"gopkg.in/redis.v4"
|
"gopkg.in/redis.v4"
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("Redis ring", func() {
|
var _ = Describe("Redis Ring", func() {
|
||||||
const heartbeat = 100 * time.Millisecond
|
const heartbeat = 100 * time.Millisecond
|
||||||
|
|
||||||
var ring *redis.Ring
|
var ring *redis.Ring
|
||||||
|
|
||||||
setRingKeys := func() {
|
setRingKeys := func() {
|
||||||
|
@ -23,20 +24,14 @@ var _ = Describe("Redis ring", func() {
|
||||||
}
|
}
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
ring = redis.NewRing(&redis.RingOptions{
|
opt := redisRingOptions()
|
||||||
Addrs: map[string]string{
|
opt.HeartbeatFrequency = heartbeat
|
||||||
"ringShardOne": ":" + ringShard1Port,
|
ring = redis.NewRing(opt)
|
||||||
"ringShardTwo": ":" + ringShard2Port,
|
|
||||||
},
|
err := ring.ForEachShard(func(cl *redis.Client) error {
|
||||||
HeartbeatFrequency: heartbeat,
|
return cl.FlushDb().Err()
|
||||||
})
|
})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
// Shards should not have any keys.
|
|
||||||
Expect(ringShard1.FlushDb().Err()).NotTo(HaveOccurred())
|
|
||||||
Expect(ringShard1.Info().Val()).NotTo(ContainSubstring("keys="))
|
|
||||||
|
|
||||||
Expect(ringShard2.FlushDb().Err()).NotTo(HaveOccurred())
|
|
||||||
Expect(ringShard2.Info().Val()).NotTo(ContainSubstring("keys="))
|
|
||||||
})
|
})
|
||||||
|
|
||||||
AfterEach(func() {
|
AfterEach(func() {
|
||||||
|
|
Loading…
Reference in New Issue