2015-01-24 15:12:48 +03:00
|
|
|
package redis_test
|
|
|
|
|
|
|
|
import (
|
2015-11-14 16:54:16 +03:00
|
|
|
"fmt"
|
2015-01-24 15:12:48 +03:00
|
|
|
"math/rand"
|
2015-04-28 18:14:19 +03:00
|
|
|
"net"
|
2015-12-16 17:11:52 +03:00
|
|
|
"strconv"
|
2015-11-14 16:54:16 +03:00
|
|
|
"strings"
|
2015-12-16 17:11:52 +03:00
|
|
|
"sync"
|
2015-04-28 18:14:19 +03:00
|
|
|
|
|
|
|
"testing"
|
2015-03-18 13:41:24 +03:00
|
|
|
"time"
|
2015-01-24 15:12:48 +03:00
|
|
|
|
|
|
|
. "github.com/onsi/ginkgo"
|
|
|
|
. "github.com/onsi/gomega"
|
2015-03-18 13:41:24 +03:00
|
|
|
|
2016-04-09 13:27:16 +03:00
|
|
|
"gopkg.in/redis.v4"
|
|
|
|
"gopkg.in/redis.v4/internal/hashtag"
|
2015-01-24 15:12:48 +03:00
|
|
|
)
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
type clusterScenario struct {
|
|
|
|
ports []string
|
|
|
|
nodeIds []string
|
|
|
|
processes map[string]*redisProcess
|
|
|
|
clients map[string]*redis.Client
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *clusterScenario) primary() *redis.Client {
|
|
|
|
return s.clients[s.ports[0]]
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *clusterScenario) masters() []*redis.Client {
|
|
|
|
result := make([]*redis.Client, 3)
|
|
|
|
for pos, port := range s.ports[:3] {
|
|
|
|
result[pos] = s.clients[port]
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
return result
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
func (s *clusterScenario) slaves() []*redis.Client {
|
|
|
|
result := make([]*redis.Client, 3)
|
|
|
|
for pos, port := range s.ports[3:] {
|
|
|
|
result[pos] = s.clients[port]
|
|
|
|
}
|
|
|
|
return result
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-05-10 16:01:38 +03:00
|
|
|
func (s *clusterScenario) clusterClient(opt *redis.ClusterOptions) *redis.ClusterClient {
|
2015-04-28 18:14:19 +03:00
|
|
|
addrs := make([]string, len(s.ports))
|
|
|
|
for i, port := range s.ports {
|
|
|
|
addrs[i] = net.JoinHostPort("127.0.0.1", port)
|
|
|
|
}
|
2015-05-10 16:01:38 +03:00
|
|
|
if opt == nil {
|
2016-03-17 19:00:47 +03:00
|
|
|
opt = &redis.ClusterOptions{
|
|
|
|
DialTimeout: 10 * time.Second,
|
|
|
|
ReadTimeout: 30 * time.Second,
|
|
|
|
WriteTimeout: 30 * time.Second,
|
|
|
|
PoolSize: 10,
|
|
|
|
PoolTimeout: 30 * time.Second,
|
|
|
|
IdleTimeout: time.Second,
|
|
|
|
IdleCheckFrequency: time.Second,
|
|
|
|
}
|
2015-05-10 16:01:38 +03:00
|
|
|
}
|
|
|
|
opt.Addrs = addrs
|
|
|
|
return redis.NewClusterClient(opt)
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
func startCluster(scenario *clusterScenario) error {
|
2015-11-14 16:54:16 +03:00
|
|
|
// Start processes and collect node ids
|
2015-04-28 18:14:19 +03:00
|
|
|
for pos, port := range scenario.ports {
|
|
|
|
process, err := startRedis(port, "--cluster-enabled", "yes")
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2016-03-16 17:57:24 +03:00
|
|
|
client := redis.NewClient(&redis.Options{
|
|
|
|
Addr: ":" + port,
|
|
|
|
})
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
info, err := client.ClusterNodes().Result()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
scenario.processes[port] = process
|
|
|
|
scenario.clients[port] = client
|
|
|
|
scenario.nodeIds[pos] = info[:40]
|
|
|
|
}
|
|
|
|
|
|
|
|
// Meet cluster nodes
|
|
|
|
for _, client := range scenario.clients {
|
|
|
|
err := client.ClusterMeet("127.0.0.1", scenario.ports[0]).Err()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
// Bootstrap masters
|
|
|
|
slots := []int{0, 5000, 10000, 16384}
|
2015-11-14 16:54:16 +03:00
|
|
|
for pos, master := range scenario.masters() {
|
|
|
|
err := master.ClusterAddSlotsRange(slots[pos], slots[pos+1]-1).Err()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
// Bootstrap slaves
|
2015-11-14 16:54:16 +03:00
|
|
|
for idx, slave := range scenario.slaves() {
|
|
|
|
masterId := scenario.nodeIds[idx]
|
|
|
|
|
|
|
|
// Wait until master is available
|
|
|
|
err := eventually(func() error {
|
|
|
|
s := slave.ClusterNodes().Val()
|
|
|
|
wanted := masterId
|
|
|
|
if !strings.Contains(s, wanted) {
|
|
|
|
return fmt.Errorf("%q does not contain %q", s, wanted)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, 10*time.Second)
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2015-11-14 16:54:16 +03:00
|
|
|
err = slave.ClusterReplicate(masterId).Err()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-11-14 16:54:16 +03:00
|
|
|
// Wait until all nodes have consistent info
|
2015-04-28 18:14:19 +03:00
|
|
|
for _, client := range scenario.clients {
|
2015-11-14 16:54:16 +03:00
|
|
|
err := eventually(func() error {
|
2015-11-22 15:44:38 +03:00
|
|
|
res, err := client.ClusterSlots().Result()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-11-21 14:16:13 +03:00
|
|
|
}
|
2016-04-09 12:52:43 +03:00
|
|
|
wanted := []redis.ClusterSlot{
|
|
|
|
{0, 4999, []redis.ClusterNode{{"", "127.0.0.1:8220"}, {"", "127.0.0.1:8223"}}},
|
|
|
|
{5000, 9999, []redis.ClusterNode{{"", "127.0.0.1:8221"}, {"", "127.0.0.1:8224"}}},
|
|
|
|
{10000, 16383, []redis.ClusterNode{{"", "127.0.0.1:8222"}, {"", "127.0.0.1:8225"}}},
|
2015-11-22 15:44:38 +03:00
|
|
|
}
|
2016-04-09 12:52:43 +03:00
|
|
|
return assertSlotsEqual(res, wanted)
|
2016-03-14 17:51:46 +03:00
|
|
|
}, 30*time.Second)
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-04-09 12:52:43 +03:00
|
|
|
func assertSlotsEqual(slots, wanted []redis.ClusterSlot) error {
|
|
|
|
outer_loop:
|
|
|
|
for _, s2 := range wanted {
|
|
|
|
for _, s1 := range slots {
|
|
|
|
if slotEqual(s1, s2) {
|
|
|
|
continue outer_loop
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return fmt.Errorf("%v not found in %v", s2, slots)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func slotEqual(s1, s2 redis.ClusterSlot) bool {
|
|
|
|
if s1.Start != s2.Start {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if s1.End != s2.End {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for i, n1 := range s1.Nodes {
|
|
|
|
if n1.Addr != s2.Nodes[i].Addr {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
func stopCluster(scenario *clusterScenario) error {
|
|
|
|
for _, client := range scenario.clients {
|
|
|
|
if err := client.Close(); err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
|
|
|
for _, process := range scenario.processes {
|
|
|
|
if err := process.Close(); err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
var _ = Describe("Cluster", func() {
|
2015-01-24 15:12:48 +03:00
|
|
|
Describe("HashSlot", func() {
|
|
|
|
|
|
|
|
It("should calculate hash slots", func() {
|
|
|
|
tests := []struct {
|
|
|
|
key string
|
|
|
|
slot int
|
|
|
|
}{
|
|
|
|
{"123456789", 12739},
|
|
|
|
{"{}foo", 9500},
|
|
|
|
{"foo{}", 5542},
|
|
|
|
{"foo{}{bar}", 8363},
|
|
|
|
{"", 10503},
|
|
|
|
{"", 5176},
|
|
|
|
{string([]byte{83, 153, 134, 118, 229, 214, 244, 75, 140, 37, 215, 215}), 5463},
|
|
|
|
}
|
2016-04-06 14:01:08 +03:00
|
|
|
// Empty keys receive random slot.
|
2015-01-24 15:12:48 +03:00
|
|
|
rand.Seed(100)
|
|
|
|
|
|
|
|
for _, test := range tests {
|
2015-12-30 16:53:45 +03:00
|
|
|
Expect(hashtag.Slot(test.key)).To(Equal(test.slot), "for %s", test.key)
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should extract keys from tags", func() {
|
|
|
|
tests := []struct {
|
|
|
|
one, two string
|
|
|
|
}{
|
|
|
|
{"foo{bar}", "bar"},
|
|
|
|
{"{foo}bar", "foo"},
|
|
|
|
{"{user1000}.following", "{user1000}.followers"},
|
|
|
|
{"foo{{bar}}zap", "{bar"},
|
|
|
|
{"foo{bar}{zap}", "bar"},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, test := range tests {
|
2015-12-30 16:53:45 +03:00
|
|
|
Expect(hashtag.Slot(test.one)).To(Equal(hashtag.Slot(test.two)), "for %s <-> %s", test.one, test.two)
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
})
|
|
|
|
|
|
|
|
})
|
|
|
|
|
|
|
|
Describe("Commands", func() {
|
|
|
|
|
|
|
|
It("should CLUSTER SLOTS", func() {
|
2015-05-14 16:13:45 +03:00
|
|
|
res, err := cluster.primary().ClusterSlots().Result()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(HaveLen(3))
|
2016-04-09 12:52:43 +03:00
|
|
|
|
|
|
|
wanted := []redis.ClusterSlot{
|
|
|
|
{0, 4999, []redis.ClusterNode{{"", "127.0.0.1:8220"}, {"", "127.0.0.1:8223"}}},
|
|
|
|
{5000, 9999, []redis.ClusterNode{{"", "127.0.0.1:8221"}, {"", "127.0.0.1:8224"}}},
|
|
|
|
{10000, 16383, []redis.ClusterNode{{"", "127.0.0.1:8222"}, {"", "127.0.0.1:8225"}}},
|
|
|
|
}
|
|
|
|
Expect(assertSlotsEqual(res, wanted)).NotTo(HaveOccurred())
|
2015-01-24 15:12:48 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER NODES", func() {
|
2015-05-14 16:13:45 +03:00
|
|
|
res, err := cluster.primary().ClusterNodes().Result()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(len(res)).To(BeNumerically(">", 400))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER INFO", func() {
|
2015-05-14 16:13:45 +03:00
|
|
|
res, err := cluster.primary().ClusterInfo().Result()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(ContainSubstring("cluster_known_nodes:6"))
|
|
|
|
})
|
|
|
|
|
2015-12-28 19:58:04 +03:00
|
|
|
It("should CLUSTER KEYSLOT", func() {
|
2015-12-29 19:16:14 +03:00
|
|
|
hashSlot, err := cluster.primary().ClusterKeySlot("somekey").Result()
|
2015-12-28 19:58:04 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2015-12-30 16:53:45 +03:00
|
|
|
Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey"))))
|
2015-12-29 19:16:14 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER COUNT-FAILURE-REPORTS", func() {
|
|
|
|
n, err := cluster.primary().ClusterCountFailureReports(cluster.nodeIds[0]).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(n).To(Equal(int64(0)))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER COUNTKEYSINSLOT", func() {
|
|
|
|
n, err := cluster.primary().ClusterCountKeysInSlot(10).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(n).To(Equal(int64(0)))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER DELSLOTS", func() {
|
|
|
|
res, err := cluster.primary().ClusterDelSlotsRange(16000, 16384-1).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(Equal("OK"))
|
|
|
|
cluster.primary().ClusterAddSlotsRange(16000, 16384-1)
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER SAVECONFIG", func() {
|
|
|
|
res, err := cluster.primary().ClusterSaveConfig().Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(Equal("OK"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER SLAVES", func() {
|
|
|
|
nodesList, err := cluster.primary().ClusterSlaves(cluster.nodeIds[0]).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(nodesList).Should(ContainElement(ContainSubstring("slave")))
|
|
|
|
Expect(nodesList).Should(HaveLen(1))
|
|
|
|
})
|
|
|
|
|
2016-06-05 12:45:39 +03:00
|
|
|
// It("should CLUSTER READONLY", func() {
|
|
|
|
// res, err := cluster.primary().ReadOnly().Result()
|
|
|
|
// Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Expect(res).To(Equal("OK"))
|
|
|
|
// })
|
|
|
|
|
|
|
|
// It("should CLUSTER READWRITE", func() {
|
|
|
|
// res, err := cluster.primary().ReadWrite().Result()
|
|
|
|
// Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Expect(res).To(Equal("OK"))
|
|
|
|
// })
|
2015-01-24 15:12:48 +03:00
|
|
|
})
|
2016-06-05 14:30:56 +03:00
|
|
|
})
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2016-06-05 14:30:56 +03:00
|
|
|
var _ = Describe("ClusterClient", func() {
|
|
|
|
var client *redis.ClusterClient
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2016-06-05 14:30:56 +03:00
|
|
|
describeClusterClient := func() {
|
2015-01-24 15:12:48 +03:00
|
|
|
It("should GET/SET/DEL", func() {
|
|
|
|
val, err := client.Get("A").Result()
|
|
|
|
Expect(err).To(Equal(redis.Nil))
|
|
|
|
Expect(val).To(Equal(""))
|
|
|
|
|
2015-01-31 12:08:56 +03:00
|
|
|
val, err = client.Set("A", "VALUE", 0).Result()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(val).To(Equal("OK"))
|
|
|
|
|
|
|
|
val, err = client.Get("A").Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(val).To(Equal("VALUE"))
|
|
|
|
|
|
|
|
cnt, err := client.Del("A").Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(cnt).To(Equal(int64(1)))
|
|
|
|
})
|
|
|
|
|
2016-06-17 15:09:38 +03:00
|
|
|
It("returns pool stats", func() {
|
2016-01-19 19:36:40 +03:00
|
|
|
Expect(client.PoolStats()).To(BeAssignableToTypeOf(&redis.PoolStats{}))
|
|
|
|
})
|
|
|
|
|
2016-06-17 15:09:38 +03:00
|
|
|
It("follows redirects", func() {
|
2015-01-31 12:08:56 +03:00
|
|
|
Expect(client.Set("A", "VALUE", 0).Err()).NotTo(HaveOccurred())
|
2015-05-01 10:42:58 +03:00
|
|
|
|
2015-12-30 16:53:45 +03:00
|
|
|
slot := hashtag.Slot("A")
|
2016-05-06 21:12:31 +03:00
|
|
|
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
|
2015-01-24 15:12:48 +03:00
|
|
|
|
|
|
|
val, err := client.Get("A").Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(val).To(Equal("VALUE"))
|
2015-03-18 13:41:24 +03:00
|
|
|
})
|
|
|
|
|
2016-06-17 15:09:38 +03:00
|
|
|
It("returns an error when there are no attempts left", func() {
|
2016-06-05 14:30:56 +03:00
|
|
|
client := cluster.clusterClient(&redis.ClusterOptions{
|
2015-05-10 16:01:38 +03:00
|
|
|
MaxRedirects: -1,
|
|
|
|
})
|
|
|
|
|
2015-12-30 16:53:45 +03:00
|
|
|
slot := hashtag.Slot("A")
|
2016-05-06 21:12:31 +03:00
|
|
|
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
|
2015-05-10 16:01:38 +03:00
|
|
|
|
|
|
|
err := client.Get("A").Err()
|
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.Error()).To(ContainSubstring("MOVED"))
|
2016-06-05 14:30:56 +03:00
|
|
|
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
2015-05-10 16:01:38 +03:00
|
|
|
})
|
2015-12-16 17:11:52 +03:00
|
|
|
|
2016-06-17 15:09:38 +03:00
|
|
|
It("supports Watch", func() {
|
2015-12-16 17:11:52 +03:00
|
|
|
var incr func(string) error
|
|
|
|
|
|
|
|
// Transactionally increments key using GET and SET commands.
|
|
|
|
incr = func(key string) error {
|
2016-05-02 15:54:15 +03:00
|
|
|
err := client.Watch(func(tx *redis.Tx) error {
|
|
|
|
n, err := tx.Get(key).Int64()
|
|
|
|
if err != nil && err != redis.Nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = tx.MultiExec(func() error {
|
|
|
|
tx.Set(key, strconv.FormatInt(n+1, 10), 0)
|
|
|
|
return nil
|
|
|
|
})
|
2015-12-16 17:11:52 +03:00
|
|
|
return err
|
2016-05-02 15:54:15 +03:00
|
|
|
}, key)
|
2015-12-16 17:11:52 +03:00
|
|
|
if err == redis.TxFailedErr {
|
|
|
|
return incr(key)
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for i := 0; i < 100; i++ {
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
2016-07-02 15:52:10 +03:00
|
|
|
defer GinkgoRecover()
|
2015-12-16 17:11:52 +03:00
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
err := incr("key")
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
n, err := client.Get("key").Int64()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(n).To(Equal(int64(100)))
|
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-06-05 14:30:56 +03:00
|
|
|
It("supports pipeline", func() {
|
2016-04-06 14:01:08 +03:00
|
|
|
slot := hashtag.Slot("A")
|
2016-05-06 21:12:31 +03:00
|
|
|
Expect(client.SwapSlotNodes(slot)).To(Equal([]string{"127.0.0.1:8224", "127.0.0.1:8221"}))
|
2016-04-06 14:01:08 +03:00
|
|
|
|
|
|
|
pipe := client.Pipeline()
|
|
|
|
defer pipe.Close()
|
|
|
|
|
|
|
|
keys := []string{"A", "B", "C", "D", "E", "F", "G"}
|
2016-06-05 14:30:56 +03:00
|
|
|
|
2016-04-06 14:01:08 +03:00
|
|
|
for i, key := range keys {
|
|
|
|
pipe.Set(key, key+"_value", 0)
|
|
|
|
pipe.Expire(key, time.Duration(i+1)*time.Hour)
|
|
|
|
}
|
2016-06-05 14:30:56 +03:00
|
|
|
cmds, err := pipe.Exec()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(cmds).To(HaveLen(14))
|
|
|
|
|
2016-04-06 14:01:08 +03:00
|
|
|
for _, key := range keys {
|
|
|
|
pipe.Get(key)
|
|
|
|
pipe.TTL(key)
|
|
|
|
}
|
2016-06-05 14:30:56 +03:00
|
|
|
cmds, err = pipe.Exec()
|
2016-04-06 14:01:08 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2016-06-05 14:30:56 +03:00
|
|
|
Expect(cmds).To(HaveLen(14))
|
|
|
|
Expect(cmds[0].(*redis.StringCmd).Val()).To(Equal("A_value"))
|
|
|
|
Expect(cmds[1].(*redis.DurationCmd).Val()).To(BeNumerically("~", 1*time.Hour, time.Second))
|
|
|
|
Expect(cmds[6].(*redis.StringCmd).Val()).To(Equal("D_value"))
|
|
|
|
Expect(cmds[7].(*redis.DurationCmd).Val()).To(BeNumerically("~", 4*time.Hour, time.Second))
|
|
|
|
Expect(cmds[12].(*redis.StringCmd).Val()).To(Equal("G_value"))
|
|
|
|
Expect(cmds[13].(*redis.DurationCmd).Val()).To(BeNumerically("~", 7*time.Hour, time.Second))
|
2016-04-06 14:01:08 +03:00
|
|
|
})
|
|
|
|
|
2016-06-05 14:30:56 +03:00
|
|
|
It("supports pipeline with missing keys", func() {
|
2016-04-06 14:01:08 +03:00
|
|
|
Expect(client.Set("A", "A_value", 0).Err()).NotTo(HaveOccurred())
|
|
|
|
Expect(client.Set("C", "C_value", 0).Err()).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
var a, b, c *redis.StringCmd
|
2016-04-09 10:47:15 +03:00
|
|
|
cmds, err := client.Pipelined(func(pipe *redis.Pipeline) error {
|
2016-04-06 14:01:08 +03:00
|
|
|
a = pipe.Get("A")
|
|
|
|
b = pipe.Get("B")
|
|
|
|
c = pipe.Get("C")
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).To(Equal(redis.Nil))
|
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
|
|
|
|
Expect(a.Err()).NotTo(HaveOccurred())
|
|
|
|
Expect(a.Val()).To(Equal("A_value"))
|
|
|
|
|
|
|
|
Expect(b.Err()).To(Equal(redis.Nil))
|
|
|
|
Expect(b.Val()).To(Equal(""))
|
|
|
|
|
|
|
|
Expect(c.Err()).NotTo(HaveOccurred())
|
|
|
|
Expect(c.Val()).To(Equal("C_value"))
|
|
|
|
})
|
2016-06-17 15:09:38 +03:00
|
|
|
|
|
|
|
It("calls fn for every master node", func() {
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
Expect(client.Set(strconv.Itoa(i), "", 0).Err()).NotTo(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
|
|
|
err := client.ForEachMaster(func(master *redis.Client) error {
|
|
|
|
return master.FlushDb().Err()
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
for _, client := range cluster.masters() {
|
|
|
|
keys, err := client.Keys("*").Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(keys).To(HaveLen(0))
|
|
|
|
}
|
|
|
|
})
|
2016-06-05 14:30:56 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
Describe("default ClusterClient", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
client = cluster.clusterClient(nil)
|
2016-06-17 15:09:38 +03:00
|
|
|
|
|
|
|
_ = client.ForEachMaster(func(master *redis.Client) error {
|
|
|
|
return master.FlushDb().Err()
|
|
|
|
})
|
2016-06-05 14:30:56 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
describeClusterClient()
|
|
|
|
})
|
|
|
|
|
|
|
|
Describe("ClusterClient with RouteByLatency", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
client = cluster.clusterClient(&redis.ClusterOptions{
|
|
|
|
RouteByLatency: true,
|
|
|
|
})
|
2016-06-17 15:09:38 +03:00
|
|
|
|
|
|
|
_ = client.ForEachMaster(func(master *redis.Client) error {
|
|
|
|
return master.FlushDb().Err()
|
|
|
|
})
|
2016-06-05 14:30:56 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
2016-06-17 15:09:38 +03:00
|
|
|
client.FlushDb()
|
2016-06-05 14:30:56 +03:00
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
describeClusterClient()
|
2016-04-06 14:01:08 +03:00
|
|
|
})
|
2016-09-23 14:52:19 +03:00
|
|
|
|
|
|
|
Describe("ClusterClient without valid nodes", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
client = redis.NewClusterClient(&redis.ClusterOptions{
|
|
|
|
Addrs: []string{redisAddr},
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
It("returns an error", func() {
|
|
|
|
err := client.Ping().Err()
|
|
|
|
Expect(err).To(MatchError("ERR This instance has cluster support disabled"))
|
|
|
|
})
|
|
|
|
})
|
2015-01-24 15:12:48 +03:00
|
|
|
})
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
//------------------------------------------------------------------------------
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
func BenchmarkRedisClusterPing(b *testing.B) {
|
2015-07-18 11:22:42 +03:00
|
|
|
if testing.Short() {
|
|
|
|
b.Skip("skipping in short mode")
|
|
|
|
}
|
|
|
|
|
2015-05-14 16:13:45 +03:00
|
|
|
cluster := &clusterScenario{
|
2015-04-28 18:14:19 +03:00
|
|
|
ports: []string{"8220", "8221", "8222", "8223", "8224", "8225"},
|
|
|
|
nodeIds: make([]string, 6),
|
|
|
|
processes: make(map[string]*redisProcess, 6),
|
|
|
|
clients: make(map[string]*redis.Client, 6),
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-05-14 16:13:45 +03:00
|
|
|
if err := startCluster(cluster); err != nil {
|
2015-04-28 18:14:19 +03:00
|
|
|
b.Fatal(err)
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-05-14 16:13:45 +03:00
|
|
|
defer stopCluster(cluster)
|
|
|
|
client := cluster.clusterClient(nil)
|
2015-04-28 18:14:19 +03:00
|
|
|
defer client.Close()
|
|
|
|
|
|
|
|
b.ResetTimer()
|
|
|
|
|
|
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
|
|
for pb.Next() {
|
|
|
|
if err := client.Ping().Err(); err != nil {
|
|
|
|
b.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|