2015-01-24 15:12:48 +03:00
|
|
|
package redis_test
|
|
|
|
|
|
|
|
import (
|
2019-07-04 11:18:06 +03:00
|
|
|
"context"
|
2015-11-14 16:54:16 +03:00
|
|
|
"fmt"
|
2015-04-28 18:14:19 +03:00
|
|
|
"net"
|
2015-12-16 17:11:52 +03:00
|
|
|
"strconv"
|
2015-11-14 16:54:16 +03:00
|
|
|
"strings"
|
2015-12-16 17:11:52 +03:00
|
|
|
"sync"
|
2015-03-18 13:41:24 +03:00
|
|
|
"time"
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2016-12-16 17:26:48 +03:00
|
|
|
. "github.com/onsi/ginkgo"
|
|
|
|
. "github.com/onsi/gomega"
|
chore: sync master (#2051)
* Upgrade redis-server version (#1833)
* Upgrade redis-server version
Signed-off-by: monkey <golang@88.com>
* XAutoClaim changed the return value
Signed-off-by: monkey <golang@88.com>
* add cmd: geosearch, geosearchstore (#1836)
* add cmd: geosearch, geosearchstore
Signed-off-by: monkey92t <golang@88.com>
* GeoSearchQuery and GeoSearchLocationQuery changed to pointer passing
Signed-off-by: monkey92t <golang@88.com>
* Added missing method XInfoStreamFull to Cmdable interface
* Run go mod tidy in redisotel
Signed-off-by: Bogdan Drutu <bogdandrutu@gmail.com>
* Revert "ConnPool check fd for bad conns (#1824)" (#1849)
This reverts commit 346bfafddd36dd52d51b064033048de5552ee91e.
* Automate release process (#1852)
* Bump github.com/onsi/gomega from 1.10.5 to 1.14.0 (#1832)
* Bump github.com/onsi/gomega from 1.10.5 to 1.14.0
Bumps [github.com/onsi/gomega](https://github.com/onsi/gomega) from 1.10.5 to 1.14.0.
- [Release notes](https://github.com/onsi/gomega/releases)
- [Changelog](https://github.com/onsi/gomega/blob/master/CHANGELOG.md)
- [Commits](https://github.com/onsi/gomega/compare/v1.10.5...v1.14.0)
---
updated-dependencies:
- dependency-name: github.com/onsi/gomega
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
* Upgrade gomega to v1.15.0
Signed-off-by: monkey92t <golang@88.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: monkey92t <golang@88.com>
* Add version.go
* Fix otel example
* Fix package name in release script
* More fixes for otel example
* And more
* Fix release.sh
* Release v8.11.3 (release.sh)
* Create an annotated tag to give release.yml chance to run
* Tweak tag.sh
* Add Cmd.Slice helper to cast to []interface{} (#1859)
* after the connection pool is closed, no new connections should be added (#1863)
* after the connection pool is closed, no new connections should be added
Signed-off-by: monkey92t <golang@88.com>
* remove runGoroutine
Signed-off-by: monkey92t <golang@88.com>
* pool.popIdle add p.closed check
Signed-off-by: monkey92t <golang@88.com>
* upgrade golangci-lint v1.42.0
Signed-off-by: monkey92t <golang@88.com>
* Bump github.com/onsi/gomega from 1.15.0 to 1.16.0 (#1865)
Bumps [github.com/onsi/gomega](https://github.com/onsi/gomega) from 1.15.0 to 1.16.0.
- [Release notes](https://github.com/onsi/gomega/releases)
- [Changelog](https://github.com/onsi/gomega/blob/master/CHANGELOG.md)
- [Commits](https://github.com/onsi/gomega/compare/v1.15.0...v1.16.0)
---
updated-dependencies:
- dependency-name: github.com/onsi/gomega
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
* Add go 1.17 to the build matrix
* Remove go 1.15 from build matrix
* Add scan struct example (#1870)
* Replace release job
* Bump github.com/cespare/xxhash/v2 from 2.1.1 to 2.1.2 (#1872)
Bumps [github.com/cespare/xxhash/v2](https://github.com/cespare/xxhash) from 2.1.1 to 2.1.2.
- [Release notes](https://github.com/cespare/xxhash/releases)
- [Commits](https://github.com/cespare/xxhash/compare/v2.1.1...v2.1.2)
---
updated-dependencies:
- dependency-name: github.com/cespare/xxhash/v2
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
* Fix tag script to push tag by tag
* Fix releasing.md
* Fix/pubsub ping mutex (#1878)
* Fix PubSub.Ping to hold the lock
* Fix PubSub.Ping to hold the lock
* add write cmd data-race test
Signed-off-by: monkey92t <golang@88.com>
Co-authored-by: monkey92t <golang@88.com>
* chore: cleanup OpenTelemetry example
* chore: gofmt all code
* Refactor TestParseURL
This is in preparation for supporting query parameters
in ParseURL:
- use an expected *Options instance to execute assertions on
- extract assertions into helper function
- enable parallel testing
- condense test table
* Add query parameter parsing to ParseURL()
Before this change, ParseURL would only accept a very restricted
set of URLs (it returned an error, if it encountered any parameter).
This commit introduces the ability to process URLs like
redis://localhost/1?dial_timeout=10s
and similar.
Go programs which were providing a configuration tunable (e.g.
CLI flag, config entry or environment variable) to configure
the Redis connection now don't need to perform this task
themselves.
* chore: add links to readme
* chore: fix discussions link
* empty hooks.withContext removed
* chore: gofmt
* chore: use conventional commits and auto-generate changelog
* feat: add acl auth support for sentinels
* chore: swap to acl auth at the test-level
* Add support for BLMove command
* chore: update dependencies
* chore: update link
* feat: add SetVal method for each command
* feat: add Cmd.{String,Int,Float,Bool}Slice helpers and an example
* chore: tweak GH actions to run all jobs
* chore: add Lua scripting example
* Fix Redis Cluster issue during roll outs of new nodes with same addr (#1914)
* fix: recycle connections in some Redis Cluster scenarios
This issue was surfaced in a Cloud Provider solution that used for
rolling out new nodes using the same address (hostname) of the nodes
that will be replaced in a Redis Cluster, while the former ones once
depromoted as Slaves would continue in service during some mintues
for redirecting traffic.
The solution basically identifies when the connection could be stale
since a MOVED response will be returned using the same address (hostname)
that is being used by the connection. At that moment we consider the
connection as no longer usable forcing to recycle the connection.
* chore: lazy reload when moved or ask
* chore: use conv commit message
* chore: release v8.11.4 (release.sh)
* fix: add whitespace for avoid unlikely colisions
* fix: format
* chore: fix links
* chore: use ctx parameter in cmdInfo
* Bump github.com/onsi/ginkgo from 1.16.4 to 1.16.5 (#1925)
Bumps [github.com/onsi/ginkgo](https://github.com/onsi/ginkgo) from 1.16.4 to 1.16.5.
- [Release notes](https://github.com/onsi/ginkgo/releases)
- [Changelog](https://github.com/onsi/ginkgo/blob/master/CHANGELOG.md)
- [Commits](https://github.com/onsi/ginkgo/compare/v1.16.4...v1.16.5)
---
updated-dependencies:
- dependency-name: github.com/onsi/ginkgo
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
* feat: add support for time.Duration write and scan
* test: add test case for setting and scanning durations
* chore: fix linter
* fix(extra/redisotel): set span.kind attribute to client
According to the opentelemetry specification this should always be set to client for database client
libraries.
I've also removed the SetAttributes call and instead set the attributes during creation of the span.
This is what the library SHOULD be doing according to the opentelemetry api specification.
* chore: update otel example
* fix: update some argument counts in pre-allocs
In some cases number of pre-allocated places in
argument array is missing 1 or 2 elements,
which results in re-allocation of twice as large array
* chore: add example how to delete keys without a ttl
* chore: don't enable all lints
* chore(deps): bump github.com/onsi/gomega from 1.16.0 to 1.17.0
Bumps [github.com/onsi/gomega](https://github.com/onsi/gomega) from 1.16.0 to 1.17.0.
- [Release notes](https://github.com/onsi/gomega/releases)
- [Changelog](https://github.com/onsi/gomega/blob/master/CHANGELOG.md)
- [Commits](https://github.com/onsi/gomega/compare/v1.16.0...v1.17.0)
---
updated-dependencies:
- dependency-name: github.com/onsi/gomega
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
* feat: Add redis v7's NX, XX, GT, LT expire variants
* chore: add missing readme
* chore: tweak feature links
* chore: remove Discord
* fix: set timeout for WAIT command. Fixes #1963
* build: update `go` directive in `go.mod` to 1.17
This commit enables support for module graph pruning and lazy module
loading for projects that are at Go 1.17 or higher.
Reference: https://go.dev/ref/mod#go-mod-file-go
Reference: https://go.dev/ref/mod#graph-pruning
Reference: https://go.dev/ref/mod#lazy-loading
Signed-off-by: Eng Zer Jun <engzerjun@gmail.com>
* chore: update link
* chore: export cmder.SetFirstKeyPos to support build module commands
* feat(redisotel): ability to override TracerProvider (#1998)
* fix: add missing Expire methods to Cmdable
This is a followup to https://github.com/go-redis/redis/pull/1928
* chore(deps): bump github.com/onsi/gomega from 1.17.0 to 1.18.1
Bumps [github.com/onsi/gomega](https://github.com/onsi/gomega) from 1.17.0 to 1.18.1.
- [Release notes](https://github.com/onsi/gomega/releases)
- [Changelog](https://github.com/onsi/gomega/blob/master/CHANGELOG.md)
- [Commits](https://github.com/onsi/gomega/compare/v1.17.0...v1.18.1)
---
updated-dependencies:
- dependency-name: github.com/onsi/gomega
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
* Update README.md (#2011)
chore: add fmt library in example code
* chore: instrumentation name and version (#2012)
* fix: invalid type assert in stringArg
* chore: cleanup
* fix: example/otel compile error (#2028)
* fix: rename Golang to Go (#2030)
https://go.dev/doc/faq#go_or_golang
* feat: add support for passing extra attributes added to spans
* feat: set net.peer.name and net.peer.port in otel example
* chore: tweak Uptrace copy
* feat: add support for COPY command (#2016)
* feat: add support for acl sentinel auth in universal client
* chore(deps): bump actions/checkout from 2 to 3
Bumps [actions/checkout](https://github.com/actions/checkout) from 2 to 3.
- [Release notes](https://github.com/actions/checkout/releases)
- [Changelog](https://github.com/actions/checkout/blob/main/CHANGELOG.md)
- [Commits](https://github.com/actions/checkout/compare/v2...v3)
---
updated-dependencies:
- dependency-name: actions/checkout
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com>
* chore: add hll example
* chore: tweak release script
* chore: release v8.11.5 (release.sh)
* chore: add discord back
Co-authored-by: Eugene Ponizovsky <ponizovsky@gmail.com>
Co-authored-by: Bogdan Drutu <bogdandrutu@gmail.com>
Co-authored-by: Vladimir Mihailenco <vladimir.webdev@gmail.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: Kishan B <kishancs46@gmail.com>
Co-authored-by: Dominik Menke <dom@digineo.de>
Co-authored-by: Gökhan Özeloğlu <gozeloglu@gmail.com>
Co-authored-by: Justin Sievenpiper <justin@sievenpiper.co>
Co-authored-by: Алексей Романовский <aromanovsky@epiphan.com>
Co-authored-by: Stavros Panakakakis <stavrospanakakis@gmail.com>
Co-authored-by: Pau Freixes <pfreixes@gmail.com>
Co-authored-by: Ethan Hur <ethan0311@gmail.com>
Co-authored-by: Jackie <18378976+Pyrodash@users.noreply.github.com>
Co-authored-by: Kristinn Björgvin Árdal <kristinnardalsecondary@gmail.com>
Co-authored-by: ffenix113 <razerer@bigmir.net>
Co-authored-by: Bastien Penavayre <bastienPenava@gmail.com>
Co-authored-by: James3 Li(李麒傑) <james3_li@asus.com>
Co-authored-by: Eng Zer Jun <engzerjun@gmail.com>
Co-authored-by: gzjiangtao2014 <gzjiangtao2014@corp.netease.com>
Co-authored-by: Nelz <nelz9999@users.noreply.github.com>
Co-authored-by: Daniel Richter <Nexyz9@gmail.com>
Co-authored-by: Seyed Ali Ghaffari <ali.ghaffari@outlook.com>
Co-authored-by: lintanghui <lintanghui@bilibili.com>
Co-authored-by: hidu <duv123+github@gmail.com>
Co-authored-by: Jonas Lergell <jonas.lergell@volvocars.com>
Co-authored-by: Alex Kahn <alexanderkahn@gmail.com>
2022-03-19 07:40:31 +03:00
|
|
|
|
2022-06-04 17:39:21 +03:00
|
|
|
"github.com/go-redis/redis/v9"
|
|
|
|
"github.com/go-redis/redis/v9/internal/hashtag"
|
2015-01-24 15:12:48 +03:00
|
|
|
)
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
type clusterScenario struct {
|
|
|
|
ports []string
|
2019-07-25 13:53:00 +03:00
|
|
|
nodeIDs []string
|
2015-04-28 18:14:19 +03:00
|
|
|
processes map[string]*redisProcess
|
|
|
|
clients map[string]*redis.Client
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *clusterScenario) masters() []*redis.Client {
|
|
|
|
result := make([]*redis.Client, 3)
|
|
|
|
for pos, port := range s.ports[:3] {
|
|
|
|
result[pos] = s.clients[port]
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
return result
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
func (s *clusterScenario) slaves() []*redis.Client {
|
|
|
|
result := make([]*redis.Client, 3)
|
|
|
|
for pos, port := range s.ports[3:] {
|
|
|
|
result[pos] = s.clients[port]
|
|
|
|
}
|
|
|
|
return result
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2017-02-17 13:12:06 +03:00
|
|
|
func (s *clusterScenario) addrs() []string {
|
2015-04-28 18:14:19 +03:00
|
|
|
addrs := make([]string, len(s.ports))
|
|
|
|
for i, port := range s.ports {
|
|
|
|
addrs[i] = net.JoinHostPort("127.0.0.1", port)
|
|
|
|
}
|
2017-02-17 13:12:06 +03:00
|
|
|
return addrs
|
|
|
|
}
|
|
|
|
|
2020-09-05 11:34:37 +03:00
|
|
|
func (s *clusterScenario) newClusterClientUnstable(opt *redis.ClusterOptions) *redis.ClusterClient {
|
2017-02-17 13:12:06 +03:00
|
|
|
opt.Addrs = s.addrs()
|
2018-12-13 13:26:02 +03:00
|
|
|
return redis.NewClusterClient(opt)
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
func (s *clusterScenario) newClusterClient(
|
|
|
|
ctx context.Context, opt *redis.ClusterOptions,
|
|
|
|
) *redis.ClusterClient {
|
2020-09-05 11:34:37 +03:00
|
|
|
client := s.newClusterClientUnstable(opt)
|
2018-07-23 12:01:13 +03:00
|
|
|
|
2018-07-22 10:50:26 +03:00
|
|
|
err := eventually(func() error {
|
2018-07-23 12:01:13 +03:00
|
|
|
if opt.ClusterSlots != nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
state, err := client.LoadState(ctx)
|
2018-05-17 16:09:56 +03:00
|
|
|
if err != nil {
|
2018-07-22 10:50:26 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-07-23 12:01:13 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
if !state.IsConsistent(ctx) {
|
2018-11-24 14:16:21 +03:00
|
|
|
return fmt.Errorf("cluster state is not consistent")
|
2018-07-23 12:01:13 +03:00
|
|
|
}
|
|
|
|
|
2018-07-22 10:50:26 +03:00
|
|
|
return nil
|
|
|
|
}, 30*time.Second)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
2018-07-23 12:01:13 +03:00
|
|
|
|
2018-05-17 16:09:56 +03:00
|
|
|
return client
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2020-09-09 17:39:13 +03:00
|
|
|
func (s *clusterScenario) Close() error {
|
|
|
|
for _, port := range s.ports {
|
|
|
|
processes[port].Close()
|
|
|
|
delete(processes, port)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
func startCluster(ctx context.Context, scenario *clusterScenario) error {
|
2015-11-14 16:54:16 +03:00
|
|
|
// Start processes and collect node ids
|
2015-04-28 18:14:19 +03:00
|
|
|
for pos, port := range scenario.ports {
|
|
|
|
process, err := startRedis(port, "--cluster-enabled", "yes")
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2016-03-16 17:57:24 +03:00
|
|
|
client := redis.NewClient(&redis.Options{
|
|
|
|
Addr: ":" + port,
|
|
|
|
})
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
info, err := client.ClusterNodes(ctx).Result()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
scenario.processes[port] = process
|
|
|
|
scenario.clients[port] = client
|
2019-07-25 13:53:00 +03:00
|
|
|
scenario.nodeIDs[pos] = info[:40]
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
// Meet cluster nodes.
|
2015-04-28 18:14:19 +03:00
|
|
|
for _, client := range scenario.clients {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.ClusterMeet(ctx, "127.0.0.1", scenario.ports[0]).Err()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
// Bootstrap masters.
|
2015-04-28 18:14:19 +03:00
|
|
|
slots := []int{0, 5000, 10000, 16384}
|
2015-11-14 16:54:16 +03:00
|
|
|
for pos, master := range scenario.masters() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := master.ClusterAddSlotsRange(ctx, slots[pos], slots[pos+1]-1).Err()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
// Bootstrap slaves.
|
2015-11-14 16:54:16 +03:00
|
|
|
for idx, slave := range scenario.slaves() {
|
2019-07-25 13:53:00 +03:00
|
|
|
masterID := scenario.nodeIDs[idx]
|
2015-11-14 16:54:16 +03:00
|
|
|
|
|
|
|
// Wait until master is available
|
|
|
|
err := eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
s := slave.ClusterNodes(ctx).Val()
|
2019-07-25 13:53:00 +03:00
|
|
|
wanted := masterID
|
2015-11-14 16:54:16 +03:00
|
|
|
if !strings.Contains(s, wanted) {
|
|
|
|
return fmt.Errorf("%q does not contain %q", s, wanted)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, 10*time.Second)
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = slave.ClusterReplicate(ctx, masterID).Err()
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-01-24 15:12:48 +03:00
|
|
|
}
|
2015-04-28 18:14:19 +03:00
|
|
|
}
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
// Wait until all nodes have consistent info.
|
2018-05-17 16:09:56 +03:00
|
|
|
wanted := []redis.ClusterSlot{{
|
|
|
|
Start: 0,
|
|
|
|
End: 4999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8220",
|
|
|
|
}, {
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8223",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 5000,
|
|
|
|
End: 9999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8221",
|
|
|
|
}, {
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8224",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 10000,
|
|
|
|
End: 16383,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8222",
|
|
|
|
}, {
|
2019-07-25 13:53:00 +03:00
|
|
|
ID: "",
|
2018-05-17 16:09:56 +03:00
|
|
|
Addr: "127.0.0.1:8225",
|
|
|
|
}},
|
|
|
|
}}
|
2015-04-28 18:14:19 +03:00
|
|
|
for _, client := range scenario.clients {
|
2015-11-14 16:54:16 +03:00
|
|
|
err := eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
res, err := client.ClusterSlots(ctx).Result()
|
2015-11-22 15:44:38 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2015-11-21 14:16:13 +03:00
|
|
|
}
|
2016-04-09 12:52:43 +03:00
|
|
|
return assertSlotsEqual(res, wanted)
|
2016-03-14 17:51:46 +03:00
|
|
|
}, 30*time.Second)
|
2015-04-28 18:14:19 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-04-09 12:52:43 +03:00
|
|
|
func assertSlotsEqual(slots, wanted []redis.ClusterSlot) error {
|
2018-10-11 13:58:31 +03:00
|
|
|
outerLoop:
|
2016-04-09 12:52:43 +03:00
|
|
|
for _, s2 := range wanted {
|
|
|
|
for _, s1 := range slots {
|
|
|
|
if slotEqual(s1, s2) {
|
2018-10-11 13:58:31 +03:00
|
|
|
continue outerLoop
|
2016-04-09 12:52:43 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return fmt.Errorf("%v not found in %v", s2, slots)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func slotEqual(s1, s2 redis.ClusterSlot) bool {
|
|
|
|
if s1.Start != s2.Start {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if s1.End != s2.End {
|
|
|
|
return false
|
|
|
|
}
|
2016-12-16 17:26:48 +03:00
|
|
|
if len(s1.Nodes) != len(s2.Nodes) {
|
|
|
|
return false
|
|
|
|
}
|
2016-04-09 12:52:43 +03:00
|
|
|
for i, n1 := range s1.Nodes {
|
|
|
|
if n1.Addr != s2.Nodes[i].Addr {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2015-04-28 18:14:19 +03:00
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
2016-10-02 15:44:01 +03:00
|
|
|
var _ = Describe("ClusterClient", func() {
|
2018-05-17 16:09:56 +03:00
|
|
|
var failover bool
|
2016-12-16 17:26:48 +03:00
|
|
|
var opt *redis.ClusterOptions
|
2016-10-02 15:44:01 +03:00
|
|
|
var client *redis.ClusterClient
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2016-12-16 17:26:48 +03:00
|
|
|
assertClusterClient := func() {
|
2019-07-04 11:18:06 +03:00
|
|
|
It("supports WithContext", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
ctx, cancel := context.WithCancel(ctx)
|
2019-07-04 11:18:06 +03:00
|
|
|
cancel()
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2019-07-04 11:18:06 +03:00
|
|
|
Expect(err).To(MatchError("context canceled"))
|
|
|
|
})
|
|
|
|
|
2015-01-24 15:12:48 +03:00
|
|
|
It("should GET/SET/DEL", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Get(ctx, "A").Err()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).To(Equal(redis.Nil))
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = client.Set(ctx, "A", "VALUE", 0).Err()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
Eventually(func() string {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.Get(ctx, "A").Val()
|
2017-08-31 15:22:47 +03:00
|
|
|
}, 30*time.Second).Should(Equal("VALUE"))
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
cnt, err := client.Del(ctx, "A").Result()
|
2015-01-24 15:12:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(cnt).To(Equal(int64(1)))
|
|
|
|
})
|
|
|
|
|
2018-05-17 16:09:56 +03:00
|
|
|
It("GET follows redirects", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Set(ctx, "A", "VALUE", 0).Err()
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2015-05-01 10:42:58 +03:00
|
|
|
|
2018-05-17 16:09:56 +03:00
|
|
|
if !failover {
|
|
|
|
Eventually(func() int64 {
|
2020-03-11 17:26:42 +03:00
|
|
|
nodes, err := client.Nodes(ctx, "A")
|
2018-05-17 16:09:56 +03:00
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
2020-03-11 17:26:42 +03:00
|
|
|
return nodes[1].Client.DBSize(ctx).Val()
|
2018-05-17 16:09:56 +03:00
|
|
|
}, 30*time.Second).Should(Equal(int64(1)))
|
2015-01-24 15:12:48 +03:00
|
|
|
|
2018-05-17 16:09:56 +03:00
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.SwapNodes(ctx, "A")
|
2018-05-17 16:09:56 +03:00
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
v, err := client.Get(ctx, "A").Result()
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(v).To(Equal("VALUE"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("SET follows redirects", func() {
|
|
|
|
if !failover {
|
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.SwapNodes(ctx, "A")
|
2018-05-17 16:09:56 +03:00
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Set(ctx, "A", "VALUE", 0).Err()
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
v, err := client.Get(ctx, "A").Result()
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(v).To(Equal("VALUE"))
|
2015-05-10 16:01:38 +03:00
|
|
|
})
|
2015-12-16 17:11:52 +03:00
|
|
|
|
2016-10-09 14:12:32 +03:00
|
|
|
It("distributes keys", func() {
|
|
|
|
for i := 0; i < 100; i++ {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Set(ctx, fmt.Sprintf("key%d", i), "value", 0).Err()
|
2016-10-09 14:12:32 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
2018-05-17 16:09:56 +03:00
|
|
|
defer GinkgoRecover()
|
2017-07-09 13:10:07 +03:00
|
|
|
Eventually(func() string {
|
2020-03-11 17:26:42 +03:00
|
|
|
return master.Info(ctx, "keyspace").Val()
|
2017-08-31 15:22:47 +03:00
|
|
|
}, 30*time.Second).Should(Or(
|
2017-07-09 13:10:07 +03:00
|
|
|
ContainSubstring("keys=31"),
|
|
|
|
ContainSubstring("keys=29"),
|
|
|
|
ContainSubstring("keys=40"),
|
|
|
|
))
|
2018-05-17 16:09:56 +03:00
|
|
|
return nil
|
|
|
|
})
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
It("distributes keys when using EVAL", func() {
|
|
|
|
script := redis.NewScript(`
|
|
|
|
local r = redis.call('SET', KEYS[1], ARGV[1])
|
|
|
|
return r
|
|
|
|
`)
|
|
|
|
|
|
|
|
var key string
|
|
|
|
for i := 0; i < 100; i++ {
|
|
|
|
key = fmt.Sprintf("key%d", i)
|
2020-03-11 17:26:42 +03:00
|
|
|
err := script.Run(ctx, client, []string{key}, "value").Err()
|
2016-10-09 14:12:32 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
2018-06-29 10:45:05 +03:00
|
|
|
defer GinkgoRecover()
|
2017-07-09 13:10:07 +03:00
|
|
|
Eventually(func() string {
|
2020-03-11 17:26:42 +03:00
|
|
|
return master.Info(ctx, "keyspace").Val()
|
2017-08-31 15:22:47 +03:00
|
|
|
}, 30*time.Second).Should(Or(
|
2017-07-09 13:10:07 +03:00
|
|
|
ContainSubstring("keys=31"),
|
|
|
|
ContainSubstring("keys=29"),
|
|
|
|
ContainSubstring("keys=40"),
|
|
|
|
))
|
2018-06-29 10:45:05 +03:00
|
|
|
return nil
|
|
|
|
})
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
|
|
|
|
2021-04-16 16:58:11 +03:00
|
|
|
It("distributes scripts when using Script Load", func() {
|
|
|
|
client.ScriptFlush(ctx)
|
|
|
|
|
|
|
|
script := redis.NewScript(`return 'Unique script'`)
|
|
|
|
|
|
|
|
script.Load(ctx, client)
|
|
|
|
|
|
|
|
client.ForEachShard(ctx, func(ctx context.Context, shard *redis.Client) error {
|
|
|
|
defer GinkgoRecover()
|
|
|
|
|
|
|
|
val, _ := script.Exists(ctx, shard).Result()
|
|
|
|
Expect(val[0]).To(Equal(true))
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
It("checks all shards when using Script Exists", func() {
|
|
|
|
client.ScriptFlush(ctx)
|
|
|
|
|
|
|
|
script := redis.NewScript(`return 'First script'`)
|
|
|
|
lostScriptSrc := `return 'Lost script'`
|
|
|
|
lostScript := redis.NewScript(lostScriptSrc)
|
|
|
|
|
|
|
|
script.Load(ctx, client)
|
|
|
|
client.Do(ctx, "script", "load", lostScriptSrc)
|
|
|
|
|
|
|
|
val, _ := client.ScriptExists(ctx, script.Hash(), lostScript.Hash()).Result()
|
|
|
|
|
|
|
|
Expect(val).To(Equal([]bool{true, false}))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("flushes scripts from all shards when using ScriptFlush", func() {
|
|
|
|
script := redis.NewScript(`return 'Unnecessary script'`)
|
|
|
|
script.Load(ctx, client)
|
|
|
|
|
|
|
|
val, _ := client.ScriptExists(ctx, script.Hash()).Result()
|
|
|
|
Expect(val).To(Equal([]bool{true}))
|
|
|
|
|
|
|
|
client.ScriptFlush(ctx)
|
|
|
|
|
|
|
|
val, _ = client.ScriptExists(ctx, script.Hash()).Result()
|
|
|
|
Expect(val).To(Equal([]bool{false}))
|
|
|
|
})
|
|
|
|
|
2016-06-17 15:09:38 +03:00
|
|
|
It("supports Watch", func() {
|
2015-12-16 17:11:52 +03:00
|
|
|
var incr func(string) error
|
|
|
|
|
|
|
|
// Transactionally increments key using GET and SET commands.
|
|
|
|
incr = func(key string) error {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Watch(ctx, func(tx *redis.Tx) error {
|
|
|
|
n, err := tx.Get(ctx, key).Int64()
|
2016-05-02 15:54:15 +03:00
|
|
|
if err != nil && err != redis.Nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err = tx.TxPipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Set(ctx, key, strconv.FormatInt(n+1, 10), 0)
|
2016-05-02 15:54:15 +03:00
|
|
|
return nil
|
|
|
|
})
|
2015-12-16 17:11:52 +03:00
|
|
|
return err
|
2016-05-02 15:54:15 +03:00
|
|
|
}, key)
|
2015-12-16 17:11:52 +03:00
|
|
|
if err == redis.TxFailedErr {
|
|
|
|
return incr(key)
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for i := 0; i < 100; i++ {
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
2016-07-02 15:52:10 +03:00
|
|
|
defer GinkgoRecover()
|
2015-12-16 17:11:52 +03:00
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
err := incr("key")
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
|
2018-02-15 14:00:54 +03:00
|
|
|
Eventually(func() string {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.Get(ctx, "key").Val()
|
2018-02-15 14:00:54 +03:00
|
|
|
}, 30*time.Second).Should(Equal("100"))
|
2015-12-16 17:11:52 +03:00
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
Describe("pipelining", func() {
|
|
|
|
var pipe *redis.Pipeline
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
assertPipeline := func() {
|
2016-12-16 17:26:48 +03:00
|
|
|
keys := []string{"A", "B", "C", "D", "E", "F", "G"}
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-12-16 17:26:48 +03:00
|
|
|
It("follows redirects", func() {
|
2018-05-17 16:09:56 +03:00
|
|
|
if !failover {
|
|
|
|
for _, key := range keys {
|
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.SwapNodes(ctx, key)
|
2018-05-17 16:09:56 +03:00
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
}
|
2016-12-16 17:26:48 +03:00
|
|
|
}
|
2016-06-05 14:30:56 +03:00
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
for i, key := range keys {
|
2020-03-11 17:26:42 +03:00
|
|
|
pipe.Set(ctx, key, key+"_value", 0)
|
|
|
|
pipe.Expire(ctx, key, time.Duration(i+1)*time.Hour)
|
2016-12-13 18:28:39 +03:00
|
|
|
}
|
2020-03-11 17:26:42 +03:00
|
|
|
cmds, err := pipe.Exec(ctx)
|
2016-12-13 18:28:39 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(cmds).To(HaveLen(14))
|
2016-06-05 14:30:56 +03:00
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
_ = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2018-02-12 17:15:40 +03:00
|
|
|
defer GinkgoRecover()
|
|
|
|
Eventually(func() int64 {
|
2020-03-11 17:26:42 +03:00
|
|
|
return node.DBSize(ctx).Val()
|
2018-02-12 17:15:40 +03:00
|
|
|
}, 30*time.Second).ShouldNot(BeZero())
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
|
2018-05-17 16:09:56 +03:00
|
|
|
if !failover {
|
|
|
|
for _, key := range keys {
|
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.SwapNodes(ctx, key)
|
2018-05-17 16:09:56 +03:00
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
}
|
2016-12-16 17:26:48 +03:00
|
|
|
}
|
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
for _, key := range keys {
|
2020-03-11 17:26:42 +03:00
|
|
|
pipe.Get(ctx, key)
|
|
|
|
pipe.TTL(ctx, key)
|
2016-12-13 18:28:39 +03:00
|
|
|
}
|
2020-03-11 17:26:42 +03:00
|
|
|
cmds, err = pipe.Exec(ctx)
|
2016-12-13 18:28:39 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(cmds).To(HaveLen(14))
|
2016-12-16 17:26:48 +03:00
|
|
|
|
|
|
|
for i, key := range keys {
|
|
|
|
get := cmds[i*2].(*redis.StringCmd)
|
|
|
|
Expect(get.Val()).To(Equal(key + "_value"))
|
|
|
|
|
|
|
|
ttl := cmds[(i*2)+1].(*redis.DurationCmd)
|
2017-08-15 10:34:05 +03:00
|
|
|
dur := time.Duration(i+1) * time.Hour
|
2018-07-23 15:55:13 +03:00
|
|
|
Expect(ttl.Val()).To(BeNumerically("~", dur, 30*time.Second))
|
2016-12-16 17:26:48 +03:00
|
|
|
}
|
2016-12-13 18:28:39 +03:00
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
It("works with missing keys", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
pipe.Set(ctx, "A", "A_value", 0)
|
|
|
|
pipe.Set(ctx, "C", "C_value", 0)
|
|
|
|
_, err := pipe.Exec(ctx)
|
2016-12-16 17:26:48 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
a := pipe.Get(ctx, "A")
|
|
|
|
b := pipe.Get(ctx, "B")
|
|
|
|
c := pipe.Get(ctx, "C")
|
|
|
|
cmds, err := pipe.Exec(ctx)
|
2016-12-13 18:28:39 +03:00
|
|
|
Expect(err).To(Equal(redis.Nil))
|
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
|
|
|
|
Expect(a.Err()).NotTo(HaveOccurred())
|
|
|
|
Expect(a.Val()).To(Equal("A_value"))
|
|
|
|
|
|
|
|
Expect(b.Err()).To(Equal(redis.Nil))
|
|
|
|
Expect(b.Val()).To(Equal(""))
|
|
|
|
|
|
|
|
Expect(c.Err()).NotTo(HaveOccurred())
|
|
|
|
Expect(c.Val()).To(Equal("C_value"))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
Describe("with Pipeline", func() {
|
2016-12-13 18:28:39 +03:00
|
|
|
BeforeEach(func() {
|
2017-05-01 18:42:58 +03:00
|
|
|
pipe = client.Pipeline().(*redis.Pipeline)
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2021-08-02 14:01:01 +03:00
|
|
|
AfterEach(func() {})
|
2016-12-13 18:28:39 +03:00
|
|
|
|
|
|
|
assertPipeline()
|
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
Describe("with TxPipeline", func() {
|
2016-12-13 18:28:39 +03:00
|
|
|
BeforeEach(func() {
|
2017-05-01 18:42:58 +03:00
|
|
|
pipe = client.TxPipeline().(*redis.Pipeline)
|
2016-12-13 18:28:39 +03:00
|
|
|
})
|
|
|
|
|
2021-08-02 14:01:01 +03:00
|
|
|
AfterEach(func() {})
|
2016-04-06 14:01:08 +03:00
|
|
|
|
2016-12-13 18:28:39 +03:00
|
|
|
assertPipeline()
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
2016-04-06 14:01:08 +03:00
|
|
|
})
|
2016-06-17 15:09:38 +03:00
|
|
|
|
2017-07-09 10:07:20 +03:00
|
|
|
It("supports PubSub", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
pubsub := client.Subscribe(ctx, "mychannel")
|
2017-07-09 10:07:20 +03:00
|
|
|
defer pubsub.Close()
|
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err := client.Publish(ctx, "mychannel", "hello").Result()
|
2017-07-09 13:10:07 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-07-09 10:07:20 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
msg, err := pubsub.ReceiveTimeout(ctx, time.Second)
|
2017-07-09 13:10:07 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-07-09 10:07:20 +03:00
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
_, ok := msg.(*redis.Message)
|
|
|
|
if !ok {
|
|
|
|
return fmt.Errorf("got %T, wanted *redis.Message", msg)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
})
|
2019-03-07 13:19:03 +03:00
|
|
|
|
|
|
|
It("supports PubSub.Ping without channels", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
pubsub := client.Subscribe(ctx)
|
2019-03-07 13:19:03 +03:00
|
|
|
defer pubsub.Close()
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := pubsub.Ping(ctx)
|
2019-03-07 13:19:03 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
2020-09-11 12:46:38 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
Describe("ClusterClient", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
opt = redisClusterOptions()
|
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
|
|
|
|
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
_ = client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
|
|
|
})
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
It("returns pool stats", func() {
|
|
|
|
stats := client.PoolStats()
|
|
|
|
Expect(stats).To(BeAssignableToTypeOf(&redis.PoolStats{}))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("returns an error when there are no attempts left", func() {
|
|
|
|
opt := redisClusterOptions()
|
|
|
|
opt.MaxRedirects = -1
|
|
|
|
client := cluster.newClusterClient(ctx, opt)
|
|
|
|
|
|
|
|
Eventually(func() error {
|
|
|
|
return client.SwapNodes(ctx, "A")
|
|
|
|
}, 30*time.Second).ShouldNot(HaveOccurred())
|
|
|
|
|
|
|
|
err := client.Get(ctx, "A").Err()
|
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.Error()).To(ContainSubstring("MOVED"))
|
|
|
|
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
It("calls fn for every master node", func() {
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
Expect(client.Set(ctx, strconv.Itoa(i), "", 0).Err()).NotTo(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
size, err := client.DBSize(ctx).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(size).To(Equal(int64(0)))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER SLOTS", func() {
|
|
|
|
res, err := client.ClusterSlots(ctx).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(HaveLen(3))
|
|
|
|
|
|
|
|
wanted := []redis.ClusterSlot{{
|
|
|
|
Start: 0,
|
|
|
|
End: 4999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8220",
|
|
|
|
}, {
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8223",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 5000,
|
|
|
|
End: 9999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8221",
|
|
|
|
}, {
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8224",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 10000,
|
|
|
|
End: 16383,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8222",
|
|
|
|
}, {
|
|
|
|
ID: "",
|
|
|
|
Addr: "127.0.0.1:8225",
|
|
|
|
}},
|
|
|
|
}}
|
|
|
|
Expect(assertSlotsEqual(res, wanted)).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER NODES", func() {
|
|
|
|
res, err := client.ClusterNodes(ctx).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(len(res)).To(BeNumerically(">", 400))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER INFO", func() {
|
|
|
|
res, err := client.ClusterInfo(ctx).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(ContainSubstring("cluster_known_nodes:6"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER KEYSLOT", func() {
|
|
|
|
hashSlot, err := client.ClusterKeySlot(ctx, "somekey").Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(hashSlot).To(Equal(int64(hashtag.Slot("somekey"))))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER GETKEYSINSLOT", func() {
|
|
|
|
keys, err := client.ClusterGetKeysInSlot(ctx, hashtag.Slot("somekey"), 1).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(len(keys)).To(Equal(0))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER COUNT-FAILURE-REPORTS", func() {
|
|
|
|
n, err := client.ClusterCountFailureReports(ctx, cluster.nodeIDs[0]).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(n).To(Equal(int64(0)))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER COUNTKEYSINSLOT", func() {
|
|
|
|
n, err := client.ClusterCountKeysInSlot(ctx, 10).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(n).To(Equal(int64(0)))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER SAVECONFIG", func() {
|
|
|
|
res, err := client.ClusterSaveConfig(ctx).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(res).To(Equal("OK"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should CLUSTER SLAVES", func() {
|
|
|
|
nodesList, err := client.ClusterSlaves(ctx, cluster.nodeIDs[0]).Result()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(nodesList).Should(ContainElement(ContainSubstring("slave")))
|
|
|
|
Expect(nodesList).Should(HaveLen(1))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should RANDOMKEY", func() {
|
|
|
|
const nkeys = 100
|
|
|
|
|
|
|
|
for i := 0; i < nkeys; i++ {
|
|
|
|
err := client.Set(ctx, fmt.Sprintf("key%d", i), "value", 0).Err()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}
|
|
|
|
|
|
|
|
var keys []string
|
|
|
|
addKey := func(key string) {
|
|
|
|
for _, k := range keys {
|
|
|
|
if k == key {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
keys = append(keys, key)
|
|
|
|
}
|
|
|
|
|
|
|
|
for i := 0; i < nkeys*10; i++ {
|
|
|
|
key := client.RandomKey(ctx).Val()
|
|
|
|
addKey(key)
|
|
|
|
}
|
|
|
|
|
|
|
|
Expect(len(keys)).To(BeNumerically("~", nkeys, nkeys/10))
|
|
|
|
})
|
2020-02-14 15:30:07 +03:00
|
|
|
|
|
|
|
It("supports Process hook", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
err = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return node.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
var stack []string
|
|
|
|
|
|
|
|
clusterHook := &hook{
|
|
|
|
beforeProcess: func(ctx context.Context, cmd redis.Cmder) (context.Context, error) {
|
|
|
|
Expect(cmd.String()).To(Equal("ping: "))
|
|
|
|
stack = append(stack, "cluster.BeforeProcess")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcess: func(ctx context.Context, cmd redis.Cmder) error {
|
|
|
|
Expect(cmd.String()).To(Equal("ping: PONG"))
|
|
|
|
stack = append(stack, "cluster.AfterProcess")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
}
|
|
|
|
client.AddHook(clusterHook)
|
|
|
|
|
2020-02-14 16:37:35 +03:00
|
|
|
nodeHook := &hook{
|
2020-02-14 15:30:07 +03:00
|
|
|
beforeProcess: func(ctx context.Context, cmd redis.Cmder) (context.Context, error) {
|
|
|
|
Expect(cmd.String()).To(Equal("ping: "))
|
|
|
|
stack = append(stack, "shard.BeforeProcess")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcess: func(ctx context.Context, cmd redis.Cmder) error {
|
|
|
|
Expect(cmd.String()).To(Equal("ping: PONG"))
|
|
|
|
stack = append(stack, "shard.AfterProcess")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
_ = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-02-14 16:37:35 +03:00
|
|
|
node.AddHook(nodeHook)
|
|
|
|
return nil
|
|
|
|
})
|
2020-02-14 15:30:07 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = client.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(stack).To(Equal([]string{
|
|
|
|
"cluster.BeforeProcess",
|
|
|
|
"shard.BeforeProcess",
|
|
|
|
"shard.AfterProcess",
|
|
|
|
"cluster.AfterProcess",
|
|
|
|
}))
|
|
|
|
|
|
|
|
clusterHook.beforeProcess = nil
|
|
|
|
clusterHook.afterProcess = nil
|
2020-02-14 16:37:35 +03:00
|
|
|
nodeHook.beforeProcess = nil
|
|
|
|
nodeHook.afterProcess = nil
|
2020-02-14 15:30:07 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
It("supports Pipeline hook", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
err = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return node.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
var stack []string
|
|
|
|
|
|
|
|
client.AddHook(&hook{
|
|
|
|
beforeProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) (context.Context, error) {
|
|
|
|
Expect(cmds).To(HaveLen(1))
|
|
|
|
Expect(cmds[0].String()).To(Equal("ping: "))
|
|
|
|
stack = append(stack, "cluster.BeforeProcessPipeline")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) error {
|
|
|
|
Expect(cmds).To(HaveLen(1))
|
|
|
|
Expect(cmds[0].String()).To(Equal("ping: PONG"))
|
|
|
|
stack = append(stack, "cluster.AfterProcessPipeline")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
})
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
_ = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-02-14 16:37:35 +03:00
|
|
|
node.AddHook(&hook{
|
2020-02-14 15:30:07 +03:00
|
|
|
beforeProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) (context.Context, error) {
|
|
|
|
Expect(cmds).To(HaveLen(1))
|
|
|
|
Expect(cmds[0].String()).To(Equal("ping: "))
|
|
|
|
stack = append(stack, "shard.BeforeProcessPipeline")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) error {
|
|
|
|
Expect(cmds).To(HaveLen(1))
|
|
|
|
Expect(cmds[0].String()).To(Equal("ping: PONG"))
|
|
|
|
stack = append(stack, "shard.AfterProcessPipeline")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
})
|
2020-02-14 16:37:35 +03:00
|
|
|
return nil
|
|
|
|
})
|
2020-02-14 15:30:07 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err = client.Pipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2020-02-14 15:30:07 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(stack).To(Equal([]string{
|
|
|
|
"cluster.BeforeProcessPipeline",
|
|
|
|
"shard.BeforeProcessPipeline",
|
|
|
|
"shard.AfterProcessPipeline",
|
|
|
|
"cluster.AfterProcessPipeline",
|
|
|
|
}))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("supports TxPipeline hook", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
err = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return node.Ping(ctx).Err()
|
2020-02-14 15:30:07 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
var stack []string
|
|
|
|
|
|
|
|
client.AddHook(&hook{
|
|
|
|
beforeProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) (context.Context, error) {
|
2021-01-09 10:27:42 +03:00
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
Expect(cmds[1].String()).To(Equal("ping: "))
|
2020-02-14 15:30:07 +03:00
|
|
|
stack = append(stack, "cluster.BeforeProcessPipeline")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) error {
|
2021-01-09 10:27:42 +03:00
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
Expect(cmds[1].String()).To(Equal("ping: PONG"))
|
2020-02-14 15:30:07 +03:00
|
|
|
stack = append(stack, "cluster.AfterProcessPipeline")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
})
|
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
_ = client.ForEachShard(ctx, func(ctx context.Context, node *redis.Client) error {
|
2020-02-14 16:37:35 +03:00
|
|
|
node.AddHook(&hook{
|
2020-02-14 15:30:07 +03:00
|
|
|
beforeProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) (context.Context, error) {
|
2020-02-14 16:37:35 +03:00
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
Expect(cmds[1].String()).To(Equal("ping: "))
|
2020-02-14 15:30:07 +03:00
|
|
|
stack = append(stack, "shard.BeforeProcessPipeline")
|
|
|
|
return ctx, nil
|
|
|
|
},
|
|
|
|
afterProcessPipeline: func(ctx context.Context, cmds []redis.Cmder) error {
|
2020-02-14 16:37:35 +03:00
|
|
|
Expect(cmds).To(HaveLen(3))
|
|
|
|
Expect(cmds[1].String()).To(Equal("ping: PONG"))
|
2020-02-14 15:30:07 +03:00
|
|
|
stack = append(stack, "shard.AfterProcessPipeline")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
})
|
2020-02-14 16:37:35 +03:00
|
|
|
return nil
|
|
|
|
})
|
2020-02-14 15:30:07 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err = client.TxPipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2020-02-14 15:30:07 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
Expect(stack).To(Equal([]string{
|
|
|
|
"cluster.BeforeProcessPipeline",
|
|
|
|
"shard.BeforeProcessPipeline",
|
|
|
|
"shard.AfterProcessPipeline",
|
|
|
|
"cluster.AfterProcessPipeline",
|
|
|
|
}))
|
|
|
|
})
|
2017-12-28 17:41:05 +03:00
|
|
|
|
2020-12-16 14:25:40 +03:00
|
|
|
It("should return correct replica for key", func() {
|
2020-12-16 18:45:06 +03:00
|
|
|
client, err := client.SlaveForKey(ctx, "test")
|
2020-12-16 14:25:40 +03:00
|
|
|
Expect(err).ToNot(HaveOccurred())
|
|
|
|
info := client.Info(ctx, "server")
|
|
|
|
Expect(info.Val()).Should(ContainSubstring("tcp_port:8224"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("should return correct master for key", func() {
|
|
|
|
client, err := client.MasterForKey(ctx, "test")
|
|
|
|
Expect(err).ToNot(HaveOccurred())
|
|
|
|
info := client.Info(ctx, "server")
|
|
|
|
Expect(info.Val()).Should(ContainSubstring("tcp_port:8221"))
|
|
|
|
})
|
|
|
|
|
2017-07-09 13:10:07 +03:00
|
|
|
assertClusterClient()
|
|
|
|
})
|
|
|
|
|
2016-06-05 14:30:56 +03:00
|
|
|
Describe("ClusterClient with RouteByLatency", func() {
|
|
|
|
BeforeEach(func() {
|
2016-12-16 17:26:48 +03:00
|
|
|
opt = redisClusterOptions()
|
2016-10-02 15:44:01 +03:00
|
|
|
opt.RouteByLatency = true
|
2020-03-11 17:26:42 +03:00
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
2016-06-17 15:09:38 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
2016-06-17 15:09:38 +03:00
|
|
|
})
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2017-08-31 15:22:47 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = client.ForEachSlave(ctx, func(ctx context.Context, slave *redis.Client) error {
|
2017-08-31 15:22:47 +03:00
|
|
|
Eventually(func() int64 {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.DBSize(ctx).Val()
|
2017-08-31 15:22:47 +03:00
|
|
|
}, 30*time.Second).Should(Equal(int64(0)))
|
|
|
|
return nil
|
|
|
|
})
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2016-06-05 14:30:56 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.ForEachSlave(ctx, func(ctx context.Context, slave *redis.Client) error {
|
|
|
|
return slave.ReadWrite(ctx).Err()
|
2017-08-31 15:22:47 +03:00
|
|
|
})
|
2018-05-17 16:09:56 +03:00
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
err = client.Close()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
2016-06-05 14:30:56 +03:00
|
|
|
})
|
|
|
|
|
2016-12-16 17:26:48 +03:00
|
|
|
assertClusterClient()
|
2016-04-06 14:01:08 +03:00
|
|
|
})
|
2018-06-29 10:45:05 +03:00
|
|
|
|
|
|
|
Describe("ClusterClient with ClusterSlots", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
failover = true
|
|
|
|
|
|
|
|
opt = redisClusterOptions()
|
2020-09-09 15:27:17 +03:00
|
|
|
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
|
2018-06-29 10:45:05 +03:00
|
|
|
slots := []redis.ClusterSlot{{
|
|
|
|
Start: 0,
|
|
|
|
End: 4999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard1Port,
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 5000,
|
|
|
|
End: 9999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard2Port,
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 10000,
|
|
|
|
End: 16383,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard3Port,
|
|
|
|
}},
|
|
|
|
}}
|
|
|
|
return slots, nil
|
|
|
|
}
|
2020-03-11 17:26:42 +03:00
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
2018-06-29 10:45:05 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
2018-06-29 10:45:05 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = client.ForEachSlave(ctx, func(ctx context.Context, slave *redis.Client) error {
|
2018-09-20 12:49:43 +03:00
|
|
|
Eventually(func() int64 {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.DBSize(ctx).Val()
|
2018-09-20 12:49:43 +03:00
|
|
|
}, 30*time.Second).Should(Equal(int64(0)))
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
failover = false
|
|
|
|
|
|
|
|
err := client.Close()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
assertClusterClient()
|
|
|
|
})
|
|
|
|
|
|
|
|
Describe("ClusterClient with RouteRandomly and ClusterSlots", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
failover = true
|
|
|
|
|
|
|
|
opt = redisClusterOptions()
|
|
|
|
opt.RouteRandomly = true
|
2020-09-09 15:27:17 +03:00
|
|
|
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
|
2018-09-20 12:49:43 +03:00
|
|
|
slots := []redis.ClusterSlot{{
|
|
|
|
Start: 0,
|
|
|
|
End: 4999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard1Port,
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 5000,
|
|
|
|
End: 9999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard2Port,
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 10000,
|
|
|
|
End: 16383,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":" + ringShard3Port,
|
|
|
|
}},
|
|
|
|
}}
|
|
|
|
return slots, nil
|
|
|
|
}
|
2020-03-11 17:26:42 +03:00
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
2018-09-20 12:49:43 +03:00
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
2018-09-20 12:49:43 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
2020-03-11 17:26:42 +03:00
|
|
|
err = client.ForEachSlave(ctx, func(ctx context.Context, slave *redis.Client) error {
|
2018-06-29 10:45:05 +03:00
|
|
|
Eventually(func() int64 {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.DBSize(ctx).Val()
|
2018-06-29 10:45:05 +03:00
|
|
|
}, 30*time.Second).Should(Equal(int64(0)))
|
2020-12-11 19:21:25 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
failover = false
|
|
|
|
|
|
|
|
err := client.Close()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
assertClusterClient()
|
|
|
|
})
|
|
|
|
|
|
|
|
Describe("ClusterClient with ClusterSlots with multiple nodes per slot", func() {
|
|
|
|
BeforeEach(func() {
|
|
|
|
failover = true
|
|
|
|
|
|
|
|
opt = redisClusterOptions()
|
|
|
|
opt.ReadOnly = true
|
|
|
|
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
|
|
|
|
slots := []redis.ClusterSlot{{
|
|
|
|
Start: 0,
|
|
|
|
End: 4999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":8220",
|
|
|
|
}, {
|
|
|
|
Addr: ":8223",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 5000,
|
|
|
|
End: 9999,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":8221",
|
|
|
|
}, {
|
|
|
|
Addr: ":8224",
|
|
|
|
}},
|
|
|
|
}, {
|
|
|
|
Start: 10000,
|
|
|
|
End: 16383,
|
|
|
|
Nodes: []redis.ClusterNode{{
|
|
|
|
Addr: ":8222",
|
|
|
|
}, {
|
|
|
|
Addr: ":8225",
|
|
|
|
}},
|
|
|
|
}}
|
|
|
|
return slots, nil
|
|
|
|
}
|
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
|
|
|
|
|
|
|
err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
|
|
|
|
return master.FlushDB(ctx).Err()
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
err = client.ForEachSlave(ctx, func(ctx context.Context, slave *redis.Client) error {
|
|
|
|
Eventually(func() int64 {
|
|
|
|
return client.DBSize(ctx).Val()
|
|
|
|
}, 30*time.Second).Should(Equal(int64(0)))
|
2018-06-29 10:45:05 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
failover = false
|
|
|
|
|
|
|
|
err := client.Close()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
assertClusterClient()
|
|
|
|
})
|
2016-12-03 18:30:13 +03:00
|
|
|
})
|
2016-09-23 14:52:19 +03:00
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
var _ = Describe("ClusterClient without nodes", func() {
|
|
|
|
var client *redis.ClusterClient
|
|
|
|
|
|
|
|
BeforeEach(func() {
|
|
|
|
client = redis.NewClusterClient(&redis.ClusterOptions{})
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
2017-08-31 15:22:47 +03:00
|
|
|
It("Ping returns an error", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(MatchError("redis: cluster has no nodes"))
|
|
|
|
})
|
|
|
|
|
|
|
|
It("pipeline returns an error", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err := client.Pipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2016-12-03 18:30:13 +03:00
|
|
|
return nil
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(MatchError("redis: cluster has no nodes"))
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
var _ = Describe("ClusterClient without valid nodes", func() {
|
|
|
|
var client *redis.ClusterClient
|
|
|
|
|
|
|
|
BeforeEach(func() {
|
|
|
|
client = redis.NewClusterClient(&redis.ClusterOptions{
|
|
|
|
Addrs: []string{redisAddr},
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
It("returns an error", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2018-02-15 14:00:54 +03:00
|
|
|
Expect(err).To(MatchError("ERR This instance has cluster support disabled"))
|
2016-12-03 18:30:13 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
It("pipeline returns an error", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err := client.Pipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2016-12-03 18:30:13 +03:00
|
|
|
return nil
|
|
|
|
})
|
2018-02-15 14:00:54 +03:00
|
|
|
Expect(err).To(MatchError("ERR This instance has cluster support disabled"))
|
2016-12-03 18:30:13 +03:00
|
|
|
})
|
|
|
|
})
|
|
|
|
|
2018-12-13 13:26:02 +03:00
|
|
|
var _ = Describe("ClusterClient with unavailable Cluster", func() {
|
|
|
|
var client *redis.ClusterClient
|
|
|
|
|
|
|
|
BeforeEach(func() {
|
|
|
|
opt := redisClusterOptions()
|
|
|
|
opt.ReadTimeout = 250 * time.Millisecond
|
|
|
|
opt.WriteTimeout = 250 * time.Millisecond
|
|
|
|
opt.MaxRedirects = 1
|
2020-09-05 11:34:37 +03:00
|
|
|
client = cluster.newClusterClientUnstable(opt)
|
2021-04-27 10:04:46 +03:00
|
|
|
Expect(client.Ping(ctx).Err()).NotTo(HaveOccurred())
|
|
|
|
|
|
|
|
for _, node := range cluster.clients {
|
|
|
|
err := node.ClientPause(ctx, 5*time.Second).Err()
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
}
|
2018-12-13 13:26:02 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
|
|
|
Expect(client.Close()).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
It("recovers when Cluster recovers", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2018-12-13 13:26:02 +03:00
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
|
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.Ping(ctx).Err()
|
2018-12-13 13:26:02 +03:00
|
|
|
}, "30s").ShouldNot(HaveOccurred())
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
var _ = Describe("ClusterClient timeout", func() {
|
|
|
|
var client *redis.ClusterClient
|
2016-10-09 14:12:32 +03:00
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
AfterEach(func() {
|
2017-03-04 14:04:27 +03:00
|
|
|
_ = client.Close()
|
2016-12-03 18:30:13 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
testTimeout := func() {
|
|
|
|
It("Ping timeouts", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Ping(ctx).Err()
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.(net.Error).Timeout()).To(BeTrue())
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
It("Pipeline timeouts", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
_, err := client.Pipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2016-10-09 14:12:32 +03:00
|
|
|
return nil
|
|
|
|
})
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.(net.Error).Timeout()).To(BeTrue())
|
2016-10-09 14:12:32 +03:00
|
|
|
})
|
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
It("Tx timeouts", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Watch(ctx, func(tx *redis.Tx) error {
|
|
|
|
return tx.Ping(ctx).Err()
|
2017-08-31 15:22:47 +03:00
|
|
|
}, "foo")
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.(net.Error).Timeout()).To(BeTrue())
|
2016-09-23 14:52:19 +03:00
|
|
|
})
|
|
|
|
|
2016-12-03 18:30:13 +03:00
|
|
|
It("Tx Pipeline timeouts", func() {
|
2020-03-11 17:26:42 +03:00
|
|
|
err := client.Watch(ctx, func(tx *redis.Tx) error {
|
|
|
|
_, err := tx.TxPipelined(ctx, func(pipe redis.Pipeliner) error {
|
|
|
|
pipe.Ping(ctx)
|
2016-12-03 18:30:13 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
return err
|
2017-08-31 15:22:47 +03:00
|
|
|
}, "foo")
|
2016-12-03 18:30:13 +03:00
|
|
|
Expect(err).To(HaveOccurred())
|
|
|
|
Expect(err.(net.Error).Timeout()).To(BeTrue())
|
2016-09-23 14:52:19 +03:00
|
|
|
})
|
2016-12-03 18:30:13 +03:00
|
|
|
}
|
2016-10-09 14:12:32 +03:00
|
|
|
|
2018-12-13 13:26:02 +03:00
|
|
|
const pause = 5 * time.Second
|
2016-12-03 18:30:13 +03:00
|
|
|
|
2017-08-31 15:22:47 +03:00
|
|
|
Context("read/write timeout", func() {
|
2016-12-03 18:30:13 +03:00
|
|
|
BeforeEach(func() {
|
|
|
|
opt := redisClusterOptions()
|
2018-06-18 12:55:26 +03:00
|
|
|
opt.ReadTimeout = 250 * time.Millisecond
|
|
|
|
opt.WriteTimeout = 250 * time.Millisecond
|
2017-08-31 15:22:47 +03:00
|
|
|
opt.MaxRedirects = 1
|
2020-03-11 17:26:42 +03:00
|
|
|
client = cluster.newClusterClient(ctx, opt)
|
2016-12-12 18:30:08 +03:00
|
|
|
|
2020-06-10 15:04:12 +03:00
|
|
|
err := client.ForEachShard(ctx, func(ctx context.Context, client *redis.Client) error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.ClientPause(ctx, pause).Err()
|
2016-12-12 18:30:08 +03:00
|
|
|
})
|
|
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
})
|
|
|
|
|
|
|
|
AfterEach(func() {
|
2020-06-10 15:04:12 +03:00
|
|
|
_ = client.ForEachShard(ctx, func(ctx context.Context, client *redis.Client) error {
|
2018-02-15 14:00:54 +03:00
|
|
|
defer GinkgoRecover()
|
2017-08-15 10:34:05 +03:00
|
|
|
Eventually(func() error {
|
2020-03-11 17:26:42 +03:00
|
|
|
return client.Ping(ctx).Err()
|
2017-08-15 10:34:05 +03:00
|
|
|
}, 2*pause).ShouldNot(HaveOccurred())
|
|
|
|
return nil
|
|
|
|
})
|
2016-12-12 18:30:08 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
testTimeout()
|
|
|
|
})
|
2015-01-24 15:12:48 +03:00
|
|
|
})
|