Redis学习笔记一

2020/12/21 Redis

# Redis 安装、启动与停止

# 解压文件
# -z或--gzip或--ungzip 通过gzip指令处理备份文件。
# -x或--extract或--get 从备份文件中还原文件。
# -v或--verbose 显示指令执行过程。
# -f<备份文件>或--file=<备份文件> 指定备份文件。
tar -zxvf redis-6.0.9.tar.gz

# 进入Redis解压目录
cd redis-6.0.9

# 编译
make

# 如果出现 “cc:命令未找到” 报错,需要安装gcc
# Centos中使用yum来安装gcc(这种方式需要能上网)
yum -y install gcc
# 若要安装g++,则输入(这种方式需要能上网)
yum -y install gcc-c++

# 如果出现 “zmalloc.h:50:31: 致命错误:jemalloc/jemalloc.h:没有那个文件或目录” 报错
make MALLOC=libc

# 继续报错,server.c:5343:176: 错误:‘struct redisServer’没有名为‘maxmemory’的成员
# 参考https://blog.csdn.net/Mr_FenKuan/article/details/111072880
yum install centos-release-scl -y
yum install devtoolset-9-gcc
scl enable devtoolset-9 bash

# 安装
make install

# 查看Redis安装位置
whereis redis-cli
whereis redis-server
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
文件 说明
redis-server Redis服务器
redis-cli Redis客户端
redis-benchmark Redis性能测试
redis-sentinel Redis哨兵,用于主从复制、分片集群,保证高可用
redis-check-rdb 检查RDB文件
redis-check-aof 检查AOF文件
# 指定配置文件启动Redis服务器
/usr/local/bin/redis-server /opt/redis/redis.conf
1
2
# 指定某端口号启动Redis客户端
/usr/local/bin/redis-cli -p 6379
1
2
# 测试连接
127.0.0.1:6379> ping
PONG
1
2
3
# 停止Redis服务
127.0.0.1:6379> shutdown
# 退出Redis客户端
not connected> exit
[root@localhost redis]# ps -ef|grep redis
root      12324  10325  0 20:41 pts/1    00:00:00 grep --color=auto redis
1
2
3
4
5
6
# 测试性能
redis-benchmark -h 127.0.0.1 -p 6379
1
2

# Redis 基本操作

# Redis有16个数据库,默认0号库,可以使用select切换库
127.0.0.1:6379> select 1
OK
127.0.0.1:6379[1]>

# 返回当前数据库中 key 的数量
dbsize

# 查找所有符合给定模式 的 key
keys *

# 检查给定 key 是否存在
exists key

# 删除给定的一个或多个 key
del key

# 设置过期时间
expire key seconds

# 查看存活时间,即还有多少秒过期
ttl key

# 以字符串的形式返回存储在 key 中的值的类型
type key

# 清空当前库
flushdb

# 清空所有库
flushall

# config get 配置项,查看配置信息
127.0.0.1:6379> config get port
1) "port"
2) "6379"
127.0.0.1:6379> config get dir
1) "dir"
2) "/opt/redis"
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39

# Redis 数据类型

# 基本数据类型

# String(字符串)

# 将键 key 的值设置为 value,如果 key 已存在,则覆盖已有值
set key value
# 将键 key 的值设置为 value,并将键 key 的生存时间设置为 seconds 秒,如果 key 已存在,则覆盖已有值
setex key seconds value
# 如果 key 不存在,则将键 key 的值设置为 value,否则什么都不做
setnx key value

# 获取
get key

# 将键 key 的值设为 value , 并返回键 key 在被设置之前的旧值
getset key value

# 批量设置
mset key value [key value]
# 批量获取
mget key [key]

# 自增1
incr key
# 增加指定步长
incrby key increment

# 自减1
decr key
# 减去指定步长
decrby key decrement
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27

# List(双向列表)

# lpush key element [element ...] 将一个或多个值依次插入到键为key的列表的头部
127.0.0.1:6379> lpush list 3 2 1 
(integer) 3

# rpush key element [element ...] 将一个或多个值依次插入到键为key的列表的底部
127.0.0.1:6379> rpush list 4 5 6
(integer) 6

# lrange key start stop 返回列表中指定区间内的元素,-1表示最后一个元素
127.0.0.1:6379> lrange list 0 -1
1) "1"
2) "2"
3) "3"
4) "4"
5) "5"
6) "6"

# llen key 返回列表元素个数
127.0.0.1:6379> llen list
(integer) 6

# lpop key [count] 用于删除并返回列表头部的元素
127.0.0.1:6379> lpop list
"1"
127.0.0.1:6379> lrange list 0 -1
1) "2"
2) "3"
3) "4"
4) "5"
5) "6"
127.0.0.1:6379> lpop list 2
1) "2"
2) "3"
127.0.0.1:6379> lrange list 0 -1
1) "4"
2) "5"
3) "6"

# rpop key [count] 用于删除并返回列表底部的元素
127.0.0.1:6379> rpop list
"6"
127.0.0.1:6379> lrange list 0 -1
1) "1"
2) "2"
3) "3"
4) "4"
5) "5"
127.0.0.1:6379> rpop list 2
1) "5"
2) "4"
127.0.0.1:6379> lrange list 0 -1
1) "1"
2) "2"
3) "3"

# linsert key BEFORE|AFTER pivot element 在指定元素前或后插入新元素
# lindex key index 获取指定索引的元素
# lpos key element [RANK rank] [COUNT num-matches] [MAXLEN len] 获取指定元素的索引
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58

# Set(无序集合)

# 添加
sadd key member [member ...]

# 随机删除成员并返回被删除的成员
spop key [count]

# 返回集合中所有成员
smembers key 
# 返回集合中随机成员
srandmember key [count]
# 判断成员是否存在
sismember key member
# 返回集合中元素个数
scard key

# 差集,返回第一个集合与其他集合之间的差异,也可以认为说第一个集合中独有的元素
sdiff key [key ...]
# 交集
sinter key [key ...]
# 并集
sunion key [key ...]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21

# Zset(有序集合)

# 添加
zadd key [NX|XX] [GT|LT] [CH] [INCR] score member [score member ...]

# 删除并返回得分最高的元素
zpopmax key [count]
# 删除并返回得分最低的元素
zpopmin key [count]

# 返回指定区间中的元素,-1表示最后一个元素
zrange key min max [BYSCORE|BYLEX] [REV] [LIMIT offset count] [WITHSCORES]
# 返回集合中元素个数
zcard key
1
2
3
4
5
6
7
8
9
10
11
12

# Hash(哈希表)

# 创建哈希表并赋值或者覆盖值
hset key field value [field value ...]
# 返回哈希表中指定字段 field 的值
hget key field

hmset key field value [field value ...]
hmget key field [field ...]

# 返回存储在 key 中哈希表的所有域
hkeys key
# 返回存储在 key 中哈希表的所有值
hvals key
# 返回哈希表中所有域和值
hgetall

# 删除返回哈希表中指定域
hdel key field [field ...]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17

# 特殊数据类型

# HyperLogLog

HyperLogLog 是用来做基数统计的算法。每个 HyperLogLog 键只需要花费 12 KB 内存,就可以计算接近 264 个不同元素的基数。

比如数据集 {1, 3, 5, 7, 5, 7, 8}, 那么这个数据集的基数集为 {1, 3, 5 ,7, 8}, 基数(不重复元素)为5。 基数估计就是在误差可接受的范围内,快速计算基数。

# 添加指定元素到 HyperLogLog 中
pfadd key element [element ...]

# 返回给定 HyperLogLog 的基数估算值
pfcount key [key ...]

# 将多个 HyperLogLog 合并为一个 HyperLogLog
pfmerge destkey sourcekey [sourcekey ...]
1
2
3
4
5
6
7
8

# GEO

GEO 主要用于存储地理位置信息,并对存储的信息进行操作,该功能在 Redis 3.2 版本新增。

# 添加地理位置的坐标
geoadd key [NX|XX] [CH] longitude latitude member [longitude latitude member ...]
# 获取地理位置的坐标
geopos key member [member ...]
# 计算两个位置之间的距离
geodist key member1 member2 [m|km|ft|mi]
# 根据用户给定的经纬度坐标来获取指定范围内的地理位置集合
georadius key longitude latitude radius m|km|ft|mi [WITHCOORD] [WITHDIST] [WITHHASH] [COUNT count [ANY]] [ASC|DESC] [STORE key] [STOREDIST key]

127.0.0.1:6379> geoadd city 116.23128 40.22077 beijing 121.48941 31.40527 shanghai 113.27324 23.15792 guangzhou 113.88308 22.55329 shenzhen
(integer) 4
127.0.0.1:6379> geopos city shenzhen
1) 1) "113.88307839632034302"
   2) "22.55329111565713873"
127.0.0.1:6379> geodist city guangzhou shenzhen km
"91.8118"
127.0.0.1:6379> georadius city 100 50 100 km
(empty array)
127.0.0.1:6379> georadius city 100 50 1000 km
(empty array)
127.0.0.1:6379> georadius city 100 50 10000 km
1) "shenzhen"
2) "guangzhou"
3) "shanghai"
4) "beijing"
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25

# BitMap

位图

# Redis 事务操作

Redis的事务不是严格意义上的事务,单个命令执行是原子性的。

一个事务从开始到执行会经历以下三个阶段:

  • 开启事务 multi
  • 命令入队
  • 执行事务 exec / 取消事务 discard

multi、exec、discard、watch 这四个指令构成了 Redis 事务处理的基础。

  • multi 用来组装一个事务;
  • exec 用来执行一个事务;
  • discard 用来取消一个事务;
  • watch 监视一个(或多个) key,一旦这些 key 在事务执行之前被改变,则取消事务的执行;
  • unwatch 取消 watch 命令对所有 key 的监视。

正常情况

# 开启事务
127.0.0.1:6379> multi
OK
# 命令入队
127.0.0.1:6379(TX)> set k1 v1
QUEUED
127.0.0.1:6379(TX)> set k2 v2
QUEUED
127.0.0.1:6379(TX)> get k1
QUEUED
127.0.0.1:6379(TX)> get k2
QUEUED
127.0.0.1:6379(TX)> set k3 v3
QUEUED
# 执行事务
127.0.0.1:6379(TX)> exec
1) OK
2) OK
3) "v1"
4) "v2"
5) OK
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21

编译时异常(语法错误,命令都不执行)

127.0.0.1:6379> multi 
OK
127.0.0.1:6379(TX)> set k1 v1
QUEUED
127.0.0.1:6379(TX)> set k2 v2
QUEUED
127.0.0.1:6379(TX)> suibianshuru666
(error) ERR unknown command `suibianshuru666`, with args beginning with: 
127.0.0.1:6379(TX)> set k3 v3
QUEUED
127.0.0.1:6379(TX)> exec
(error) EXECABORT Transaction discarded because of previous errors.

1
2
3
4
5
6
7
8
9
10
11
12
13

运行时异常(其他错误,命令部分执行)

127.0.0.1:6379> multi
OK
127.0.0.1:6379(TX)> set k1 v1
QUEUED
127.0.0.1:6379(TX)> set k2 v2
QUEUED
127.0.0.1:6379(TX)> incr k1
QUEUED
127.0.0.1:6379(TX)> set k3 v3
QUEUED
127.0.0.1:6379(TX)> exec
1) OK
2) OK
3) (error) ERR value is not an integer or out of range
4) OK
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15

乐观锁

127.0.0.1:6379> watch balance # 监听余额
OK
127.0.0.1:6379> multi
OK
127.0.0.1:6379(TX)> decrby balance 10 # 扣减余额
QUEUED
127.0.0.1:6379(TX)> incrby paymeng 10 # 新增支出
QUEUED
127.0.0.1:6379(TX)> exec # 执行事务,如果在执行事务之前balance被修改,则事务执行失败
1) (integer) 9990
2) (integer) 10
1
2
3
4
5
6
7
8
9
10
11

# Redis 发布订阅

# 订阅一个或多个频道的信息
subscribe channel [channel ...]

# 将信息发送到指定频道
publish channel message

# 退订一个或多个频道
unsubscribe [channel [channel ...]]
1
2
3
4
5
6
7
8

# Redis 配置文件

################################# GENERAL #####################################
# By default Redis does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
# When Redis is supervised by upstart or systemd, this parameter has no impact.
daemonize yes # 将no改为yes
1
2
3
4
5

# Redis 持久化

Redis 提供了两种持久化的方式,分别是RDB(Redis DataBase)和AOF(Append Only File)。

  1. 默认开启RDB,关闭AOF;
  2. 如果RDB和AOF同时存在,则优先采用AOF方式,这是因为AOF方式的数据恢复完整度更高;
  3. 不管是RDB还是AOF,主线程负责接收客户端请求进行读写操作,新开启的线程负责生成RDB或AOF文件。

# RDB(Redis DataBase)

RDB,在一定的时间窗口内进行了指定次数的写操作,则将Redis存储的数据生成快照并存储到磁盘等介质中。

################################ SNAPSHOTTING  ################################

# Save the DB to disk.
#
# save <seconds> <changes>
#
# Redis will save the DB if both the given number of seconds and the given
# number of write operations against the DB occurred.
#
# Snapshotting can be completely disabled with a single empty string argument
# as in following example:
#
# save "" # 禁用RDB
#
# Unless specified otherwise, by default Redis will save the DB:
#   * After 3600 seconds (an hour) if at least 1 key changed
#   * After 300 seconds (5 minutes) if at least 100 keys changed
#   * After 60 seconds if at least 10000 keys changed
#
# You can set these explicitly by uncommenting the three following lines.
#
# save 3600 1
# save 300 100
# save 60 10000

# The filename where to dump the DB
dbfilename dump.rdb

# The working directory.
#
# The DB will be written inside this directory, with the filename specified
# above using the 'dbfilename' configuration directive.
#
# The Append Only File will also be created inside this directory.
#
# Note that you must specify a directory here, not a file name.
dir ./
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37

dump.rdb在以下情况时会生成:

  1. 满足配置文件的条件,在一定的时间窗口内进行了指定次数的写操作;
   # Unless specified otherwise, by default Redis will save the DB:
   #   * After 3600 seconds (an hour) if at least 1 key changed
   #   * After 300 seconds (5 minutes) if at least 100 keys changed
   #   * After 60 seconds if at least 10000 keys changed
   #
   # You can set these explicitly by uncommenting the three following lines.
   #
   # save 3600 1
   # save 300 100
   # save 60 10000
1
2
3
4
5
6
7
8
9
10
  1. 执行 save 命令;

  2. 执行 flushdb 命令;

  3. 执行 flushall 命令;

  4. 执行 shutdown 命令。

# AOF(Append Only File)

AOF,是将在Redis中执行的所有写指令记录下来,在下次Redis服务启动的时候,将这些写指令从前到后执行一遍,实现数据恢复。

############################## APPEND ONLY MODE ###############################

# By default Redis asynchronously dumps the dataset on disk. This mode is
# good enough in many applications, but an issue with the Redis process or
# a power outage may result into a few minutes of writes lost (depending on
# the configured save points).
#
# The Append Only File is an alternative persistence mode that provides
# much better durability. For instance using the default data fsync policy
# (see later in the config file) Redis can lose just one second of writes in a
# dramatic event like a server power outage, or a single write if something
# wrong with the Redis process itself happens, but the operating system is
# still running correctly.
#
# AOF and RDB persistence can be enabled at the same time without problems.
# If the AOF is enabled on startup Redis will load the AOF, that is the file
# with the better durability guarantees.
#
# Please check http://redis.io/topics/persistence for more information.

appendonly no # 是否开启AOF方式

# The name of the append only file (default: "appendonly.aof")

appendfilename "appendonly.aof"

# The fsync() call tells the Operating System to actually write data on disk
# instead of waiting for more data in the output buffer. Some OS will really flush
# data on disk, some other OS will just try to do it ASAP.
#
# Redis supports three different modes:
#
# no: don't fsync, just let the OS flush the data when it wants. Faster.
# always: fsync after every write to the append only log. Slow, Safest.
# everysec: fsync only one time every second. Compromise.
#
# The default is "everysec", as that's usually the right compromise between
# speed and data safety. It's up to you to understand if you can relax this to
# "no" that will let the operating system flush the output buffer when
# it wants, for better performances (but if you can live with the idea of
# some data loss consider the default persistence mode that's snapshotting),
# or on the contrary, use "always" that's very slow but a bit safer than
# everysec.
#
# More details please check the following article:
# http://antirez.com/post/redis-persistence-demystified.html
#
# If unsure, use "everysec".

# appendfsync always
appendfsync everysec
# appendfsync no
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52

# Redis 主从复制

测试主从复制:同一机器创建三个配置文件,启动三个不同端口号的Redis服务,并分别使用三个客户端连接。

配置文件如下:

redis_6379.conf

# 公用配置
include /opt/redis/redis.conf

# 个性化配置
port 6379
pidfile /var/run/redis_6379.pid
logfile "redis_6379.log"
dbfilename dump_6379.rdb
1
2
3
4
5
6
7
8

redis_6380.conf

# 公用配置
include /opt/redis/redis.conf

# 个性化配置
port 6380
pidfile /var/run/redis_6380.pid
logfile "redis_6380.log"
dbfilename dump_6380.rdb
1
2
3
4
5
6
7
8

redis_6381.conf

# 公用配置
include /opt/redis/redis.conf

# 个性化配置
port 6381
pidfile /var/run/redis_6381.pid
logfile "redis_6381.log"
dbfilename dump_6381.rdb
1
2
3
4
5
6
7
8

启动服务,连接客户端

# 启动服务器
/usr/local/bin/redis-server /opt/redis/redis_6379.conf
/usr/local/bin/redis-server /opt/redis/redis_6380.conf
/usr/local/bin/redis-server /opt/redis/redis_6381.conf

# 启动客户端
/usr/local/bin/redis-cli -p 6379
/usr/local/bin/redis-cli -p 6380
/usr/local/bin/redis-cli -p 6381
1
2
3
4
5
6
7
8
9

使用 info replication 查看主从配置,Redis服务器启动默认是master节点。

127.0.0.1:6379> info replication
# Replication
role:master
connected_slaves:0
master_failover_state:no-failover
master_replid:d8ddf17b411a6e03263cf8f870bd1c2b5fd6e732
master_replid2:0000000000000000000000000000000000000000
master_repl_offset:0
second_repl_offset:-1
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
1
2
3
4
5
6
7
8
9
10
11
12
13

使用 slaveof host port 手动绑定主从,或者在配置文件中配置 replicaof <masterip> <masterport>

127.0.0.1:6380> slaveof 127.0.0.1 6379
OK
127.0.0.1:6380> info replication
# Replication
role:slave
master_host:127.0.0.1
master_port:6379
master_link_status:up
master_last_io_seconds_ago:10
master_sync_in_progress:0
slave_repl_offset:112
slave_priority:100
slave_read_only:1
connected_slaves:0
master_failover_state:no-failover
master_replid:02496d2e289dd49a98a6f8dde2c153fb1587ba6e
master_replid2:0000000000000000000000000000000000000000
master_repl_offset:112
second_repl_offset:-1
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:1
repl_backlog_histlen:112
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
127.0.0.1:6379> info replication
# Replication
role:master
connected_slaves:2
slave0:ip=127.0.0.1,port=6380,state=online,offset=126,lag=1
slave1:ip=127.0.0.1,port=6381,state=online,offset=126,lag=1
master_failover_state:no-failover
master_replid:02496d2e289dd49a98a6f8dde2c153fb1587ba6e
master_replid2:0000000000000000000000000000000000000000
master_repl_offset:126
second_repl_offset:-1
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:1
repl_backlog_histlen:126
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15

使用 slaveof no one 解绑主从

127.0.0.1:6381> slaveof no one
OK
127.0.0.1:6381> info replication
# Replication
role:master
connected_slaves:0
master_failover_state:no-failover
master_replid:954afb72e238ca2fe9b2e84009d52585207ab7c3
master_replid2:02496d2e289dd49a98a6f8dde2c153fb1587ba6e
master_repl_offset:361
second_repl_offset:362
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:29
repl_backlog_histlen:333
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15

使用 slaveof host port 绑定主从,会有两个问题:

  1. 主服务器宕机,从服务器群龙无首;
  2. 从服务器断开,重新连接时,需要手动执行slaveof绑定主从。

为了解决以上问题,引用哨兵模式。

哨兵模式,是从服务器经过投票选举,票数高的自动提升为主服务器。

sentinel.conf

# By default Redis Sentinel does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis-sentinel.pid when
# daemonized.
daemonize yes # 以守护进程启动哨兵

# sentinel monitor <master-name> <ip> <redis-port> <quorum>
#
# Tells Sentinel to monitor this master, and to consider it in O_DOWN
# (Objectively Down) state only if at least <quorum> sentinels agree.
#
# Note that whatever is the ODOWN quorum, a Sentinel will require to
# be elected by the majority of the known Sentinels in order to
# start a failover, so no failover can be performed in minority.
#
# Replicas are auto-discovered, so you don't need to specify replicas in
# any way. Sentinel itself will rewrite this configuration file adding
# the replicas using additional configuration options.
# Also note that the configuration file is rewritten when a
# replica is promoted to master.
#
# Note: master name should not include special characters or spaces.
# The valid charset is A-z 0-9 and the three characters ".-_".
sentinel monitor mymaster 127.0.0.1 6379 1 # 1表示执行故障恢复操作前至少需要几个哨兵节点同意。因为只有两个从节点,因此这个数值(quorum)只能设置为1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23

启动哨兵

/usr/local/bin/redis-sentinel /opt/redis/sentinel.conf
1

关掉6379,查看哨兵日志

[root@localhost redis]# /usr/local/bin/redis-sentinel /opt/redis/sentinel.conf
6189:X 21 Mar 2021 12:04:21.330 # oO0OoO0OoO0Oo Redis is starting oO0OoO0OoO0Oo
6189:X 21 Mar 2021 12:04:21.331 # Redis version=6.2.1, bits=64, commit=00000000, modified=0, pid=6189, just started
6189:X 21 Mar 2021 12:04:21.331 # Configuration loaded
6189:X 21 Mar 2021 12:04:21.331 * Increased maximum number of open files to 10032 (it was originally set to 1024).
6189:X 21 Mar 2021 12:04:21.331 * monotonic clock: POSIX clock_gettime
                _._                                                  
           _.-``__ ''-._                                             
      _.-``    `.  `_.  ''-._           Redis 6.2.1 (00000000/0) 64 bit
  .-`` .-```.  ```\/    _.,_ ''-._                                   
 (    '      ,       .-`  | `,    )     Running in sentinel mode
 |`-._`-...-` __...-.``-._|'` _.-'|     Port: 26379
 |    `-._   `._    /     _.-'    |     PID: 6189
  `-._    `-._  `-./  _.-'    _.-'                                   
 |`-._`-._    `-.__.-'    _.-'_.-'|                                  
 |    `-._`-._        _.-'_.-'    |           http://redis.io        
  `-._    `-._`-.__.-'_.-'    _.-'                                   
 |`-._`-._    `-.__.-'    _.-'_.-'|                                  
 |    `-._`-._        _.-'_.-'    |                                  
  `-._    `-._`-.__.-'_.-'    _.-'                                   
      `-._    `-.__.-'    _.-'                                       
          `-._        _.-'                                           
              `-.__.-'                                               

6189:X 21 Mar 2021 12:04:21.332 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.
6189:X 21 Mar 2021 12:04:21.334 # Sentinel ID is 2fecb2124682385875472f734a46504ae74760fa
6189:X 21 Mar 2021 12:04:21.334 # +monitor master mymaster 127.0.0.1 6379 quorum 1
6189:X 21 Mar 2021 12:04:21.335 * +slave slave 127.0.0.1:6380 127.0.0.1 6380 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:04:21.337 * +slave slave 127.0.0.1:6381 127.0.0.1 6381 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:14.936 # +sdown master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:14.936 # +odown master mymaster 127.0.0.1 6379 #quorum 1/1
6189:X 21 Mar 2021 12:05:14.936 # +new-epoch 1
6189:X 21 Mar 2021 12:05:14.936 # +try-failover master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:14.941 # +vote-for-leader 2fecb2124682385875472f734a46504ae74760fa 1
6189:X 21 Mar 2021 12:05:14.941 # +elected-leader master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:14.941 # +failover-state-select-slave master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.018 # +selected-slave slave 127.0.0.1:6380 127.0.0.1 6380 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.018 * +failover-state-send-slaveof-noone slave 127.0.0.1:6380 127.0.0.1 6380 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.118 * +failover-state-wait-promotion slave 127.0.0.1:6380 127.0.0.1 6380 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.331 # +promoted-slave slave 127.0.0.1:6380 127.0.0.1 6380 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.331 # +failover-state-reconf-slaves master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:15.361 * +slave-reconf-sent slave 127.0.0.1:6381 127.0.0.1 6381 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:16.325 * +slave-reconf-inprog slave 127.0.0.1:6381 127.0.0.1 6381 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:16.325 * +slave-reconf-done slave 127.0.0.1:6381 127.0.0.1 6381 @ mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:16.425 # +failover-end master mymaster 127.0.0.1 6379
6189:X 21 Mar 2021 12:05:16.425 # +switch-master mymaster 127.0.0.1 6379 127.0.0.1 6380
6189:X 21 Mar 2021 12:05:16.425 * +slave slave 127.0.0.1:6381 127.0.0.1 6381 @ mymaster 127.0.0.1 6380
6189:X 21 Mar 2021 12:05:16.425 * +slave slave 127.0.0.1:6379 127.0.0.1 6379 @ mymaster 127.0.0.1 6380
6189:X 21 Mar 2021 12:05:46.468 # +sdown slave 127.0.0.1:6379 127.0.0.1 6379 @ mymaster 127.0.0.1 6380
6189:X 21 Mar 2021 12:06:09.740 # -sdown slave 127.0.0.1:6379 127.0.0.1 6379 @ mymaster 127.0.0.1 6380
6189:X 21 Mar 2021 12:06:19.731 * +convert-to-slave slave 127.0.0.1:6379 127.0.0.1 6379 @ mymaster 127.0.0.1 6380
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51

# Redis 集群

Redis 集群,数据自动在多个Redis节点间分片。

Redis集群提供的能力:

  1. 自动切分数据集到多个节点上。
  2. 当部分节点故障或不可达的情况下继续提供服务。
# 启动服务
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6379.conf
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6380.conf
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6381.conf
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6389.conf
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6390.conf
[root@localhost redis]# /usr/local/bin/redis-server /opt/redis/cluster/redis_6391.conf

[root@localhost redis]# ps -ef|grep redis
root       8545      1  0 13:39 ?        00:00:00 /usr/local/bin/redis-server *:6379 [cluster]
root       8550      1  0 13:39 ?        00:00:00 /usr/local/bin/redis-server *:6380 [cluster]
root       8567      1  0 13:40 ?        00:00:00 /usr/local/bin/redis-server *:6381 [cluster]
root       8572      1  0 13:40 ?        00:00:00 /usr/local/bin/redis-server *:6389 [cluster]
root       8577      1  0 13:40 ?        00:00:00 /usr/local/bin/redis-server *:6390 [cluster]
root       8582      1  0 13:40 ?        00:00:00 /usr/local/bin/redis-server *:6391 [cluster]
root       8672   2228  0 13:40 pts/0    00:00:00 grep --color=auto redis

# 创建集群,选项 --cluster-replicas 1 表示为每一个主服务器配一个从服务器
# ./redis-cli --cluster create ${IP}:${PORT} ${IP}:${PORT} ${IP}:${PORT} ${IP}:${PORT} ${IP}:${PORT} ${IP}:${PORT} --cluster-replicas 1
[root@localhost redis]# /usr/local/bin/redis-cli --cluster create 192.168.66.133:6379 192.168.66.133:6380 192.168.66.133:6381 192.168.66.133:6389 192.168.66.133:6390 192.168.66.133:6391 --cluster-replicas 1
>>> Performing hash slots allocation on 6 nodes...
Master[0] -> Slots 0 - 5460
Master[1] -> Slots 5461 - 10922
Master[2] -> Slots 10923 - 16383
Adding replica 192.168.66.133:6390 to 192.168.66.133:6379
Adding replica 192.168.66.133:6391 to 192.168.66.133:6380
Adding replica 192.168.66.133:6389 to 192.168.66.133:6381
>>> Trying to optimize slaves allocation for anti-affinity
[WARNING] Some slaves are in the same host as their master
M: b8a032204ae864920827c9825bc1df5c2bf19f8f 192.168.66.133:6379
   slots:[0-5460] (5461 slots) master
M: a970ee79f6939aa6f15dbba55e7cfc32f4de193f 192.168.66.133:6380
   slots:[5461-10922] (5462 slots) master
M: 79074d9fe1f4d59bc6e027c326112094ade7bf93 192.168.66.133:6381
   slots:[10923-16383] (5461 slots) master
S: d656dba4a474c9c840c44ce90d8ef847b36fff47 192.168.66.133:6389
   replicates b8a032204ae864920827c9825bc1df5c2bf19f8f
S: 2a54c3642936c5b19751d50664811df5f87696a3 192.168.66.133:6390
   replicates a970ee79f6939aa6f15dbba55e7cfc32f4de193f
S: 78fee4677ca3aeba72e7432926f887c0e0417aa5 192.168.66.133:6391
   replicates 79074d9fe1f4d59bc6e027c326112094ade7bf93
Can I set the above configuration? (type 'yes' to accept): yes
>>> Nodes configuration updated
>>> Assign a different config epoch to each node
>>> Sending CLUSTER MEET messages to join the cluster
Waiting for the cluster to join
..
>>> Performing Cluster Check (using node 192.168.66.133:6379)
M: b8a032204ae864920827c9825bc1df5c2bf19f8f 192.168.66.133:6379
   slots:[0-5460] (5461 slots) master
   1 additional replica(s)
S: d656dba4a474c9c840c44ce90d8ef847b36fff47 192.168.66.133:6389
   slots: (0 slots) slave
   replicates b8a032204ae864920827c9825bc1df5c2bf19f8f
M: a970ee79f6939aa6f15dbba55e7cfc32f4de193f 192.168.66.133:6380
   slots:[5461-10922] (5462 slots) master
   1 additional replica(s)
M: 79074d9fe1f4d59bc6e027c326112094ade7bf93 192.168.66.133:6381
   slots:[10923-16383] (5461 slots) master
   1 additional replica(s)
S: 78fee4677ca3aeba72e7432926f887c0e0417aa5 192.168.66.133:6391
   slots: (0 slots) slave
   replicates 79074d9fe1f4d59bc6e027c326112094ade7bf93
S: 2a54c3642936c5b19751d50664811df5f87696a3 192.168.66.133:6390
   slots: (0 slots) slave
   replicates a970ee79f6939aa6f15dbba55e7cfc32f4de193f
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.

# 进入客户端,-c表示以集群方式启动
[root@localhost redis]# /usr/local/bin/redis-cli  -p 6379 -c

# 测试
127.0.0.1:6379> set k1 v1
-> Redirected to slot [12706] located at 192.168.66.133:6381
OK
192.168.66.133:6381> set k2 v2
-> Redirected to slot [449] located at 192.168.66.133:6379
OK
192.168.66.133:6379> set k3 v3
OK
192.168.66.133:6379> set k4 v4
-> Redirected to slot [8455] located at 192.168.66.133:6380
OK
192.168.66.133:6380> set k5 v5
-> Redirected to slot [12582] located at 192.168.66.133:6381
OK
192.168.66.133:6381> set k6 v6
-> Redirected to slot [325] located at 192.168.66.133:6379
OK
192.168.66.133:6379> dbsize
(integer) 3
192.168.66.133:6379> 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95

# Redis 应用

  1. 热点数据缓存
  2. 单点登录缓存用户信息
  3. 分布式锁

# Java 操作 Redis

Redis 出厂配置,默认只接受本机访问请求。

如果应用程序与 Redis 服务器不在同一台服务器,需要修改redis.conf。

  1. 将 bind 127.0.0.1 注释掉
  2. 将 protected-mode yes 改为 protected-mode no
vi /path/to/redis.conf

################################## NETWORK #####################################
# bind 127.0.0.1
protected-mode no
1
2
3
4
5

# Jedis

Jedis直连

Jedis连接池

两种方式比较

配置优化

<!-- https://mvnrepository.com/artifact/redis.clients/jedis -->
<dependency>
	<groupId>redis.clients</groupId>
	<artifactId>jedis</artifactId>
	<version>3.5.2</version>
</dependency>
1
2
3
4
5
6
// 直连
Jedis jedis = new Jedis("192.168.66.133", 6379);

// 连接池
JedisPoolConfig jedisPoolConfig = new JedisPoolConfig();
jedisPoolConfig.setMaxTotal(10);
jedisPoolConfig.setMaxWaitMillis(1000);
JedisPool jedisPool = new JedisPool(jedisPoolConfig, "192.168.66.133", 6379);
Jedis jedis = null;
try {
	jedis = jedisPool.getResource();
	System.out.println(jedis.ping());
} catch (Exception e) {
	e.printStackTrace();
} finally {
	if (jedis != null) {
		jedis.close();
	}
}
jedisPool.close();

// 集群
Set<HostAndPort> nodes = new HashSet<>();
nodes.add(new HostAndPort("192.168.66.133", 6379));
nodes.add(new HostAndPort("192.168.66.133", 6380));
nodes.add(new HostAndPort("192.168.66.133", 6381));
JedisCluster jedisCluster = new JedisCluster(nodes);
jedisCluster.close();
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28

# RedisTemplate

    <dependencyManagement>
        <dependencies>
            <dependency>
                <groupId>org.springframework.boot</groupId>
                <artifactId>spring-boot-starter-parent</artifactId>
                <version>2.3.5.RELEASE</version>
                <type>pom</type>
                <scope>import</scope>
            </dependency>
        </dependencies>
    </dependencyManagement>

    <dependencies>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-web</artifactId>
        </dependency>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-data-redis</artifactId>
        </dependency>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-test</artifactId>
        </dependency>
        <!-- Redis序列化 -->
        <dependency>
            <groupId>com.alibaba</groupId>
            <artifactId>fastjson</artifactId>
            <version>1.2.75</version>
        </dependency>
        <!-- Redis要使用连接池,需要引入commons-pool2 -->
        <dependency>
            <groupId>org.apache.commons</groupId>
            <artifactId>commons-pool2</artifactId>
        </dependency>
    </dependencies>
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
# 自动配置类:RedisAutoConfiguration
# 配置属性:RedisProperties

spring:
  redis:
    # 单机配置
#    host: 192.168.66.133
#    port: 6379
    # 集群配置
    cluster:
      nodes:
        - 192.168.66.133:6379
        - 192.168.66.133:6390
        - 192.168.66.133:6391
    # 单机或集群配置连接池,SpringBoot2.X采用Lettuce客户端连接Redis服务端,默认不使用连接池,只有配置以下属性时才使用连接池
    lettuce:
      shutdown-timeout: 100 # 关闭超时时间
      pool:
        max-active: 8 # 连接池最大连接数(使用负值表示没有限制)
        max-idle: 8   # 连接池中的最大空闲连接
        max-wait: 30  # 连接池最大阻塞等待时间(使用负值表示没有限制)
        min-idle: 0   # 连接池中的最小空闲连接
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
@Configuration
public class RedisConfig {

    @Bean
    public RedisTemplate<String, Object> redisTemplate(RedisConnectionFactory redisConnectionFactory)
            throws UnknownHostException {
        RedisTemplate<String, Object> template = new RedisTemplate<>();

        // key的序列化采用StringRedisSerializer
        StringRedisSerializer stringRedisSerializer = new StringRedisSerializer();
        template.setKeySerializer(stringRedisSerializer);
        template.setHashKeySerializer(stringRedisSerializer);

        // value的序列化采用fastJsonRedisSerializer
        FastJsonRedisSerializer fastJsonRedisSerializer = new FastJsonRedisSerializer(Object.class);
        template.setValueSerializer(fastJsonRedisSerializer);
        template.setHashValueSerializer(fastJsonRedisSerializer);

        template.setConnectionFactory(redisConnectionFactory);
        return template;
    }

    @Bean
    public StringRedisTemplate stringRedisTemplate(RedisConnectionFactory redisConnectionFactory)
            throws UnknownHostException {
        StringRedisTemplate template = new StringRedisTemplate();

        // key的序列化采用StringRedisSerializer
        StringRedisSerializer stringRedisSerializer = new StringRedisSerializer();
        template.setKeySerializer(stringRedisSerializer);

        // value的序列化采用fastJsonRedisSerializer
        FastJsonRedisSerializer fastJsonRedisSerializer = new FastJsonRedisSerializer(Object.class);
        template.setValueSerializer(fastJsonRedisSerializer);

        template.setConnectionFactory(redisConnectionFactory);
        return template;
    }

}
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
// 使 @Autowired 等注解标记的类能实例化到IOC容器中,否则会报NullPointerExecption
@RunWith(SpringRunner.class)
@SpringBootTest
public class TestRedis {

    @Autowired
    private RedisTemplate redisTemplate;

    @Test
    public void test1() {
        // 数据类型
        // redisTemplate.opsForValue()  字符串
        // redisTemplate.opsForList()   列表
        // redisTemplate.opsForSet()    无序列表
        // redisTemplate.opsForZSet()   有序列表
        // redisTemplate.opsForHash()   哈希表
        // redisTemplate.opsForHyperLogLog() HyperLogLog
        // redisTemplate.opsForGeo()    地理位置

        // key操作
        // redisTemplate.keys(pattern)
        // redisTemplate.hasKey(key)
        // redisTemplate.delete(key)

        // 事务操作
        // redisTemplate.watch(key)
        // redisTemplate.multi()
        // redisTemplate.exec()
        // redisTemplate.discard()

        // redisTemplate.expire()

        // 字符串
        redisTemplate.opsForValue().set("k1", "v1");
        redisTemplate.opsForValue().set("k2", "v2");
        redisTemplate.opsForValue().set("k3", "v3");
        redisTemplate.opsForValue().set("k4", "v4");
        redisTemplate.opsForValue().set("k5", "v5");
        redisTemplate.opsForValue().set("k6", "v6");
        redisTemplate.opsForValue().set("k7", "v7");
        redisTemplate.opsForValue().set("k8", "v8");
        redisTemplate.opsForValue().set("k9", "v9");
    }

}
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45