docker安装redis Docker安装redis docker安装Redis 详细教程_docker 安装redis(1)

Set the max number of connected clients at the same time. By default

this limit is set to 10000 clients, however if the Redis server is not

able to configure the process file limit to allow for the specified limit

the max number of allowed clients is set to the current file limit

minus 32 (as Redis reserves a few file descriptors for internal uses).

Once the limit is reached Redis will close all the new connections sending

an error ‘max number of clients reached’.

maxclients 10000

If Redis is to be used as an in-memory-only cache without any kind of

persistence, then the fork() mechanism used by the background AOF/RDB

persistence is unnecessary. As an optimization, all persistence can be

turned off in the Windows version of Redis. This will redirect heap

allocations to the system heap allocator, and disable commands that would

otherwise cause fork() operations: BGSAVE and BGREWRITEAOF.

This flag may not be combined with any of the other flags that configure

AOF and RDB operations.

persistence-available [(yes)|no]

Don’t use more memory than the specified amount of bytes.

When the memory limit is reached Redis will try to remove keys

according to the eviction policy selected (see maxmemory-policy).

If Redis can’t remove keys according to the policy, or if the policy is

set to ‘noeviction’, Redis will start to reply with errors to commands

that would use more memory, like SET, LPUSH, and so on, and will continue

to reply to read-only commands like GET.

This option is usually useful when using Redis as an LRU cache, or to set

a hard memory limit for an instance (using the ‘noeviction’ policy).

WARNING: If you have slaves attached to an instance with maxmemory on,

the size of the output buffers needed to feed the slaves are subtracted

from the used memory count, so that network problems / resyncs will

not trigger a loop where keys are evicted, and in turn the output

buffer of slaves is full with DELs of keys evicted triggering the deletion

of more keys, and so forth until the database is completely emptied.

In short… if you have slaves attached it is suggested that you set a lower

limit for maxmemory so that there is some free RAM on the system for slave

output buffers (but this is not needed if the policy is ‘noeviction’).

WARNING: not setting maxmemory will cause Redis to terminate with an

out-of-memory exception if the heap limit is reached.

NOTE: since Redis uses the system paging file to allocate the heap memory,

the Working Set memory usage showed by the Windows Task Manager or by other

tools such as ProcessExplorer will not always be accurate. For example, right

after a background save of the RDB or the AOF files, the working set value

may drop significantly. In order to check the correct amount of memory used

by the redis-server to store the data, use the INFO client command. The INFO

command shows only the memory used to store the redis data, not the extra

memory used by the Windows process for its own requirements. Th3 extra amount

of memory not reported by the INFO command can be calculated subtracting the

Peak Working Set reported by the Windows Task Manager and the used_memory_peak

reported by the INFO command.

maxmemory

MAXMEMORY POLICY: how Redis will select what to remove when maxmemory

is reached. You can select among five behaviors:

volatile-lru -> remove the key with an expire set using an LRU algorithm

allkeys-lru -> remove any key according to the LRU algorithm

volatile-random -> remove a random key with an expire set

allkeys-random -> remove a random key, any key

volatile-ttl -> remove the key with the nearest expire time (minor TTL)

noeviction -> don’t expire at all, just return an error on write operations

Note: with any of the above policies, Redis will return an error on write

operations, when there are no suitable keys for eviction.

At the date of writing these commands are: set setnx setex append

incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd

sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby

zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby

getset mset msetnx exec sort

The default is:

maxmemory-policy noeviction

LRU and minimal TTL algorithms are not precise algorithms but approximated

algorithms (in order to save memory), so you can select as well the sample

size to check. For instance for default Redis will check three keys and

pick the one that was used less recently, you can change the sample size

using the following configuration directive.

maxmemory-samples 3

############################## APPEND ONLY MODE ###############################

By default Redis asynchronously dumps the dataset on disk. This mode is

good enough in many applications, but an issue with the Redis process or

a power outage may result into a few minutes of writes lost (depending on

the configured save points).

The Append Only File is an alternative persistence mode that provides

much better durability. For instance using the default data fsync policy

(see later in the config file) Redis can lose just one second of writes in a

dramatic event like a server power outage, or a single write if something

wrong with the Redis process itself happens, but the operating system is

still running correctly.

AOF and RDB persistence can be enabled at the same time without problems.

If the AOF is enabled on startup Redis will load the AOF, that is the file

with the better durability guarantees.

Please check http://redis.io/topics/persistence for more information.

appendonly no

The name of the append only file (default: “appendonly.aof”)

appendfilename “appendonly.aof”

The fsync() call tells the Operating System to actually write data on disk

instead of waiting for more data in the output buffer. Some OS will really flush

data on disk, some other OS will just try to do it ASAP.

Redis supports three different modes:

no: don’t fsync, just let the OS flush the data when it wants. Faster.

always: fsync after every write to the append only log . Slow, Safest.

everysec: fsync only one time every second. Compromise.

The default is “everysec”, as that’s usually the right compromise between

speed and data safety. It’s up to you to understand if you can relax this to

“no” that will let the operating system flush the output buffer when

it wants, for better performances (but if you can live with the idea of

some data loss consider the default persistence mode that’s snapshotting),

or on the contrary, use “always” that’s very slow but a bit safer than

everysec.

More details please check the following article:

http://antirez.com/post/redis-persistence-demystified.html

If unsure, use “everysec”.

appendfsync always

appendfsync everysec

appendfsync no

When the AOF fsync policy is set to always or everysec, and a background

saving process (a background save or AOF log background rewriting) is

performing a lot of I/O against the disk, in some Linux configurations

Redis may block too long on the fsync() call. Note that there is no fix for

this currently, as even performing fsync in a different thread will block

our synchronous write(2) call.

In order to mitigate this problem it’s possible to use the following option

that will prevent fsync() from being called in the main process while a

BGSAVE or BGREWRITEAOF is in progress.

This means that while another child is saving, the durability of Redis is

the same as “appendfsync none”. In practical terms, this means that it is

possible to lose up to 30 seconds of log in the worst scenario (with the

default Linux settings).

If you have latency problems turn this to “yes”. Otherwise leave it as

“no” that is the safest pick from the point of view of durability.

no-appendfsync-on-rewrite no

Automatic rewrite of the append only file.

Redis is able to automatically rewrite the log file implicitly calling

BGREWRITEAOF when the AOF log size grows by the specified percentage.

This is how it works: Redis remembers the size of the AOF file after the

latest rewrite (if no rewrite has happened since the restart, the size of

the AOF at startup is used).

This base size is compared to the current size. If the current size is

bigger than the specified percentage, the rewrite is triggered. Also

you need to specify a minimal size for the AOF file to be rewritten, this

is useful to avoid rewriting the AOF file even if the percentage increase

is reached but it is still pretty small.

Specify a percentage of zero in order to disable the automatic AOF

rewrite feature.

auto-aof-rewrite-percentage 100
auto-aof-rewrite-min-size 64mb

An AOF file may be found to be truncated at the end during the Redis

startup process, when the AOF data gets loaded back into memory.

This may happen when the system where Redis is running

crashes, especially when an ext4 filesystem is mounted without the

data=ordered option (however this can’t happen when Redis itself

crashes or aborts but the operating system still works correctly).

Redis can either exit with an error when this happens, or load as much

data as possible (the default now) and start if the AOF file is found

to be truncated at the end. The following option controls this behavior.

If aof-load-truncated is set to yes, a truncated AOF file is loaded and

the Redis server starts emitting a log to inform the user of the event.

Otherwise if the option is set to no, the server aborts with an error

and refuses to start. When the option is set to no, the user requires

to fix the AOF file using the “redis-check-aof” utility before to restart

the server.

Note that if the AOF file will be found to be corrupted in the middle

the server will still exit with an error. This option only applies when

Redis will try to read more data from the AOF file but not enough bytes

will be found.

aof-load-truncated yes

################################ LUA SCRIPTING ###############################

Max execution time of a Lua script in milliseconds.

If the maximum execution time is reached Redis will log that a script is

still in execution after the maximum allowed time and will start to

reply to queries with an error.

When a long running script exceeds the maximum execution time only the

SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be

used to stop a script that did not yet called write commands. The second

is the only way to shut down the server in the case a write command was

already issued by the script but the user doesn’t want to wait for the natural

termination of the script.

Set it to 0 or a negative value for unlimited execution without warnings.

lua-time-limit 5000

################################ REDIS CLUSTER ###############################

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

WARNING EXPERIMENTAL: Redis Cluster is considered to be stable code, however

in order to mark it as “mature” we need to wait for a non trivial percentage

of users to deploy it in production.

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Normal Redis instances can’t be part of a Redis Cluster; only nodes that are

started as cluster nodes can. In order to start a Redis instance as a

cluster node enable the cluster support uncommenting the following:

cluster-enabled yes

Every cluster node has a cluster configuration file. This file is not

intended to be edited by hand. It is created and updated by Redis nodes.

Every Redis Cluster node requires a different cluster configuration file.

Make sure that instances running in the same system do not have

overlapping cluster configuration file names.

cluster-config-file nodes-6379.conf

Cluster node timeout is the amount of milliseconds a node must be unreachable

for it to be considered in failure state.

Most other internal time limits are multiple of the node timeout.

cluster-node-timeout 15000

A slave of a failing master will avoid to start a failover if its data

looks too old.

There is no simple way for a slave to actually have a exact measure of

its “data age”, so the following two checks are performed:

1) If there are multiple slaves able to failover, they exchange messages

in order to try to give an advantage to the slave with the best

replication offset (more data from the master processed).

Slaves will try to get their rank by offset, and apply to the start

of the failover a delay proportional to their rank.

2) Every single slave computes the time of the last interaction with

its master. This can be the last ping or command received (if the master

is still in the “connected” state), or the time that elapsed since the

disconnection with the master (if the replication link is currently down).

If the last interaction is too old, the slave will not try to failover

at all.

The point “2” can be tuned by user. Specifically a slave will not perform

the failover if, since the last interaction with the master, the time

elapsed is greater than:

(node-timeout * slave-validity-factor) + repl-ping-slave-period

So for example if node-timeout is 30 seconds, and the slave-validity-factor

is 10, and assuming a default repl-ping-slave-period of 10 seconds, the

slave will not try to failover if it was not able to talk with the master

for longer than 310 seconds.

A large slave-validity-factor may allow slaves with too old data to failover

a master, while a too small value may prevent the cluster from being able to

elect a slave at all.

For maximum availability, it is possible to set the slave-validity-factor

to a value of 0, which means, that slaves will always try to failover the

master regardless of the last time they interacted with the master.

(However they’ll always try to apply a delay proportional to their

offset rank).

Zero is the only value able to guarantee that when all the partitions heal

the cluster will always be able to continue.

cluster-slave-validity-factor 10

Cluster slaves are able to migrate to orphaned masters, that are masters

that are left without working slaves. This improves the cluster ability

to resist to failures as otherwise an orphaned master can’t be failed over

in case of failure if it has no working slaves.

Slaves migrate to orphaned masters only if there are still at least a

given number of other working slaves for their old master. This number

is the “migration barrier”. A migration barrier of 1 means that a slave

will migrate only if there is at least 1 other working slave for its master

and so forth. It usually reflects the number of slaves you want for every

master in your cluster.

Default is 1 (slaves migrate only if their masters remain with at least

one slave). To disable migration just set it to a very large value.

A value of 0 can be set but is useful only for debugging and dangerous

in production.

cluster-migration-barrier 1

By default Redis Cluster nodes stop accepting queries if they detect there

is at least an hash slot uncovered (no available node is serving it).

This way if the cluster is partially down (for example a range of hash slots

are no longer covered) all the cluster becomes, eventually, unavailable.

It automatically returns available as soon as all the slots are covered again.

However sometimes you want the subset of the cluster which is working,

to continue to accept queries for the part of the key space that is still

covered. In order to do so, just set the cluster-require-full-coverage

option to no.

cluster-require-full-coverage yes

In order to setup your cluster make sure to read the documentation

available at http://redis.io web site.

################################## SLOW LOG ###################################

The Redis Slow Log is a system to log queries that exceeded a specified

execution time. The execution time does not include the I/O operations

like talking with the client, sending the reply and so forth,

but just the time needed to actually execute the command (this is the only

stage of command execution where the thread is blocked and can not serve

other requests in the meantime).

You can configure the slow log with two parameters: one tells Redis

what is the execution time, in microseconds, to exceed in order for the

command to get logged, and the other parameter is the length of the

slow log. When a new command is logged the oldest one is removed from the

queue of logged commands.

The following time is expressed in microseconds, so 1000000 is equivalent

to one second. Note that a negative number disables the slow log, while

a value of zero forces the logging of every command.

slowlog-log-slower-than 10000

There is no limit to this length. Just be aware that it will consume memory.

You can reclaim memory used by the slow log with SLOWLOG RESET.

slowlog-max-len 128

################################ LATENCY MONITOR ##############################

The Redis latency monitoring subsystem samples different operations

at runtime in order to collect data related to possible sources of

latency of a Redis instance.

Via the LATENCY command this information is available to the user that can

print graphs and obtain reports.

The system only logs operations that were performed in a time equal or

greater than the amount of milliseconds specified via the

latency-monitor-threshold configuration directive. When its value is set

to zero, the latency monitor is turned off.

By default latency monitoring is disabled since it is mostly not needed

if you don’t have latency issues, and collecting data has a performance

impact, that while very small, can be measured under big load. Latency

monitoring can easily be enabled at runtime using the command

“CONFIG SET latency-monitor-threshold ” if needed.

latency-monitor-threshold 0

############################# Event notification ##############################

Redis can notify Pub/Sub clients about events happening in the key space.

This feature is documented at http://redis.io/topics/notifications

For instance if keyspace events notification is enabled, and a client

performs a DEL operation on key “foo” stored in the Database 0, two

messages will be published via Pub/Sub:

PUBLISH keyspace@0:foo del

PUBLISH keyevent@0:del foo

It is possible to select the events that Redis will notify among a set

of classes. Every class is identified by a single character:

K Keyspace events, published with keyspace@ prefix.

E Keyevent events, published with keyevent@ prefix.

g Generic commands (non-type specific) like DEL, EXPIRE, RENAME, …

$ String commands

l List commands

s Set commands

h Hash commands

z Sorted set commands

x Expired events (events generated every time a key expires)

e Evicted events (events generated when a key is evicted for maxmemory)

A Alias for g$lshzxe, so that the “AKE” string means all the events.

The “notify-keyspace-events” takes as argument a string that is composed

of zero or multiple characters. The empty string means that notifications

are disabled.

Example: to enable list and generic events, from the point of view of the

event name, use:

notify-keyspace-events Elg

Example 2: to get the stream of the expired keys subscribing to channel

name keyevent@0:expired use:

notify-keyspace-events Ex

By default all notifications are disabled because most users don’t need

this feature and the feature has some overhead. Note that if you don’t

specify at least one of K or E, no events will be delivered.

notify-keyspace-events “”

############################### ADVANCED CONFIG ###############################

Hashes are encoded using a memory efficient data structure when they have a

small number of entries, and the biggest entry does not exceed a given

threshold. These thresholds can be configured using the following directives.

hash-max-ziplist-entries 512
hash-max-ziplist-value 64

Similarly to hashes, small lists are also encoded in a special way in order

to save a lot of space. The special representation is only used when

you are under the following limits:

list-max-ziplist-entries 512
list-max-ziplist-value 64

Sets have a special encoding in just one case: when a set is composed

of just strings that happen to be integers in radix 10 in the range

of 64 bit signed integers.

The following configuration setting sets the limit in the size of the

set in order to use this special memory saving encoding.

set-max-intset-entries 512

Similarly to hashes and lists, sorted sets are also specially encoded in

order to save a lot of space. This encoding is only used when the length and

elements of a sorted set are below the following limits:

zset-max-ziplist-entries 128
zset-max-ziplist-value 64

HyperLogLog sparse representation bytes limit. The limit includes the

16 bytes header. When an HyperLogLog using the sparse representation crosses

this limit, it is converted into the dense representation.

A value greater than 16000 is totally useless, since at that point the

dense representation is more memory efficient.

The suggested value is ~ 3000 in order to have the benefits of

the space efficient encoding without slowing down too much PFADD,

which is O(N) with the sparse encoding. The value can be raised to

~ 10000 when CPU is not a concern, but space is, and the data set is

composed of many HyperLogLogs with cardinality in the 0 - 15000 range.

hll-sparse-max-bytes 3000

Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in

order to help rehashing the main Redis hash table (the one mapping top-level

keys to values). The hash table implementation Redis uses (see dict.c)

performs a lazy rehashing: the more operation you run into a hash table

that is rehashing, the more rehashing “steps” are performed, so if the

server is idle the rehashing is never complete and some more memory is used

by the hash table.

The default is to use this millisecond 10 times every second in order to

actively rehash the main dictionaries, freeing memory when possible.

If unsure:

use “activerehashing no” if you have hard latency requirements and it is

not a good thing in your environment that Redis can reply from time to time

to queries with 2 milliseconds delay.

use “activerehashing yes” if you don’t have such hard requirements but

want to free memory asap when possible.

activerehashing yes

The client output buffer limits can be used to force disconnection of clients

that are not reading data from the server fast enough for some reason (a

common reason is that a Pub/Sub client can’t consume messages as fast as the

publisher can produce them).

The limit can be set differently for the three different classes of clients:

normal -> normal clients including MONITOR clients

slave -> slave clients

pubsub -> clients subscribed to at least one pubsub channel or pattern

The syntax of every client-output-buffer-limit directive is the following:

client-output-buffer-limit

A client is immediately disconnected once the hard limit is reached, or if

the soft limit is reached and remains reached for the specified number of

seconds (continuously).

So for instance if the hard limit is 32 megabytes and the soft limit is

16 megabytes / 10 seconds, the client will get disconnected immediately

if the size of the output buffers reach 32 megabytes, but will also get

disconnected if the client reaches 16 megabytes and continuously overcomes

the limit for 10 seconds.

By default normal clients are not limited because they don’t receive data

without asking (in a push way), but just after a request, so only

asynchronous clients may create a scenario where data is requested faster

than it can read.

Instead there is a default limit for pubsub and slave clients, since

subscribers and slaves receive data in a push fashion.

Both the hard or the soft limit can be disabled by setting them to zero.

client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 256mb 64mb 60
client-output-buffer-limit pubsub 32mb 8mb 60

Redis calls an internal function to perform many background tasks, like

closing connections of clients in timeot, purging expired keys that are

never requested, and so forth.

Not all tasks are perforemd with the same frequency, but Redis checks for

tasks to perform according to the specified “hz” value.

By default “hz” is set to 10. Raising the value will use more CPU when

Redis is idle, but at the same time will make Redis more responsive when

there are many keys expiring at the same time, and timeouts may be

handled with more precision.

The range is between 1 and 500, however a value over 100 is usually not

a good idea. Most users should use the default of 10 and raise this up to

100 only in environments where very low latency is required.

hz 10

自我介绍一下,小编13年上海交大毕业,曾经在小公司待过,也去过华为、OPPO等大厂,18年进入阿里一直到现在。

深知大多数Linux运维工程师,想要提升技能,往往是自己摸索成长或者是报班学习,但对于培训机构动则几千的学费,着实压力不小。自己不成体系的自学效果低效又漫长,而且极易碰到天花板技术停滞不前!

因此收集整理了一份《2024年Linux运维全套学习资料》,初衷也很简单,就是希望能够帮助到想自学提升又不知道该从何学起的朋友,同时减轻大家的负担。
img
img
img
img
img

既有适合小白学习的零基础资料,也有适合3年以上经验的小伙伴深入学习提升的进阶课程,基本涵盖了95%以上Linux运维知识点,真正体系化!

由于文件比较大,这里只是将部分目录大纲截图出来,每个节点里面都包含大厂面经、学习笔记、源码讲义、实战项目、讲解视频,并且后续会持续更新

如果你觉得这些内容对你有帮助,可以添加VX:vip1024b (备注Linux运维获取)
img

为了做好运维面试路上的助攻手,特整理了上百道 【运维技术栈面试题集锦】 ,让你面试不慌心不跳,高薪offer怀里抱!

这次整理的面试题,小到shell、MySQL,大到K8s等云原生技术栈,不仅适合运维新人入行面试需要,还适用于想提升进阶跳槽加薪的运维朋友。

本份面试集锦涵盖了

  • 174 道运维工程师面试题
  • 128道k8s面试题
  • 108道shell脚本面试题
  • 200道Linux面试题
  • 51道docker面试题
  • 35道Jenkis面试题
  • 78道MongoDB面试题
  • 17道ansible面试题
  • 60道dubbo面试题
  • 53道kafka面试
  • 18道mysql面试题
  • 40道nginx面试题
  • 77道redis面试题
  • 28道zookeeper

总计 1000+ 道面试题, 内容 又全含金量又高

  • 174道运维工程师面试题

1、什么是运维?

2、在工作中,运维人员经常需要跟运营人员打交道,请问运营人员是做什么工作的?

3、现在给你三百台服务器,你怎么对他们进行管理?

4、简述raid0 raid1raid5二种工作模式的工作原理及特点

5、LVS、Nginx、HAproxy有什么区别?工作中你怎么选择?

6、Squid、Varinsh和Nginx有什么区别,工作中你怎么选择?

7、Tomcat和Resin有什么区别,工作中你怎么选择?

8、什么是中间件?什么是jdk?

9、讲述一下Tomcat8005、8009、8080三个端口的含义?

10、什么叫CDN?

11、什么叫网站灰度发布?

12、简述DNS进行域名解析的过程?

13、RabbitMQ是什么东西?

14、讲一下Keepalived的工作原理?

15、讲述一下LVS三种模式的工作过程?

16、mysql的innodb如何定位锁问题,mysql如何减少主从复制延迟?

17、如何重置mysql root密码?

一个人可以走的很快,但一群人才能走的更远。如果你从事以下工作或对以下感兴趣,欢迎戳这里加入程序员的圈子,让我们一起学习成长!

AI人工智能、Android移动开发、AIGC大模型、C C#、Go语言、Java、Linux运维、云计算、MySQL、PMP、网络安全、Python爬虫、UE5、UI设计、Unity3D、Web前端开发、产品经理、车载开发、大数据、鸿蒙、计算机网络、嵌入式物联网、软件测试、数据结构与算法、音视频开发、Flutter、IOS开发、PHP开发、.NET、安卓逆向、云计算

7、Tomcat和Resin有什么区别,工作中你怎么选择?

8、什么是中间件?什么是jdk?

9、讲述一下Tomcat8005、8009、8080三个端口的含义?

10、什么叫CDN?

11、什么叫网站灰度发布?

12、简述DNS进行域名解析的过程?

13、RabbitMQ是什么东西?

14、讲一下Keepalived的工作原理?

15、讲述一下LVS三种模式的工作过程?

16、mysql的innodb如何定位锁问题,mysql如何减少主从复制延迟?

17、如何重置mysql root密码?

一个人可以走的很快,但一群人才能走的更远。如果你从事以下工作或对以下感兴趣,欢迎戳这里加入程序员的圈子,让我们一起学习成长!

AI人工智能、Android移动开发、AIGC大模型、C C#、Go语言、Java、Linux运维、云计算、MySQL、PMP、网络安全、Python爬虫、UE5、UI设计、Unity3D、Web前端开发、产品经理、车载开发、大数据、鸿蒙、计算机网络、嵌入式物联网、软件测试、数据结构与算法、音视频开发、Flutter、IOS开发、PHP开发、.NET、安卓逆向、云计算

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值