配置redis主從IP和port:
master and slave | ip | port |
---|---|---|
master | 127.0.0.1 | 6379 |
slave1 | 127.0.0.1 | 6380 |
slave2 | 127.0.0.1 | 6381 |
下載 redis 源碼包
wget http://download.redis.io/releases/redis-3.2.1.tar.gz
解壓縮 并且 安裝redis
# 解壓文件
tar zxvf redis-3.2.1.tar.gz
# 修改名稱
mv redis-3.2.1 redis-3.2.1.master
# 切換目錄
cd redis-3.2.1.master/src
# 執(zhí)行make和make test
make && make test
# 執(zhí)行成功后會提示:
All tests passed without errors!
# 切換到 上級目錄
cd ..
# 復(fù)制多個 redis 實例
cp -r redis-3.2.1.master redis-3.2.1.slave1
cp -r redis-3.2.1.master redis-3.2.1.slave2
修改配置 主服務(wù)器(master)的 配置文件
# 切換目錄
cd redis-3.2.1.master
# 打開 redis.conf(配置文件)
vim redis.conf
# 修改以下配置為(根據(jù)字段修改 不是覆蓋全部) 然后保存
port 6379 # 占用的端口號
pidfile /var/run/redis_6379.pid # 當(dāng)運行多個 redis 服務(wù)時,需要指定不同的 pid 文件和端口
# slaveof <masterip> <masterport> # 當(dāng)本機為從服務(wù)時,設(shè)置主服務(wù)的IP及端口 我們這個是主服務(wù)器 不需要設(shè)置
logfile "/data/logs/redis.master.log" # log 文件存儲路徑 首先要測試該路徑是否存在
daemonize yes # Redis默認(rèn)不是以守護進(jìn)程的方式運行,可以通過該配置項修改,使用yes啟用守護進(jìn)程
requirepass password # 設(shè)置訪問密碼 密碼自行設(shè)置 一定要足夠強壯
# 切換到上級目錄
cd ..
修改配置 從服務(wù)器(slave1, slave2)的 配置文件
注意是兩臺實例都要改哦筷频,改動的流程和配置是一樣的,只不過有些配置的值不一樣柱告。
下面是改了其中一個實例截驮,另一個實例也是這么配置的。
# 切換目錄
cd redis-3.2.1.slave1
# 打開 redis.conf(配置文件)
vim redis.conf
# 修改以下配置為(根據(jù)字段修改 不是覆蓋全部) 然后保存
port 6380 # slave1 改成 6380 slave2改成 6381
pidfile /var/run/redis_6380.pid # slave1 改成 6380 slave2改成 6381
slaveof 127.0.0.1 6379 # 配置 主服務(wù)器的 IP 和 端口
masterauth password # 配置 訪問主服務(wù)器的的密碼
logfile "/data/logs/redis.master.log" # slave1 將 master 改成 6380际度, slave2改成 6381
daemonize yes # Redis默認(rèn)不是以守護進(jìn)程的方式運行葵袭,可以通過該配置項修改,使用yes啟用守護進(jìn)程
requirepass password # 設(shè)置訪問密碼 密碼自行設(shè)置 一定要足夠強壯
依次啟動 redis 實例
# 切換目錄
cd redis-3.2.1.master
# 啟動 主服務(wù)器實例
./src/redis-server redis.conf
# 切換到 slave1
cd ../redis-3.2.1.slave1
# 啟動 從服務(wù)器(slave1)實例
./src/redis-server redis.conf
# 切換到 slave2
cd ../redis-3.2.1.slave2
# 啟動 從服務(wù)器(slave1)實例
./src/redis-server redis.conf
測試主從關(guān)系
# 打開 主服務(wù)器 命令行
./redis-3.2.1.master/src/redis-cli
# 設(shè)置一個值
set test_key test
# 獲取 輸入的值
get test_key
# 退出命令行
ctrl + c
# 進(jìn)入 從服務(wù)器
./redis-3.2.1.slave1/src/redis-cli -h 127.0.0.1 -p 6380
# 獲取剛才在主服務(wù)器上設(shè)置的值
get test_key
Redis配置文件redis.conf參數(shù)說明
# By default Redis does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
#Redis默認(rèn)不是以守護進(jìn)程的方式運行乖菱,可以通過該配置項修改坡锡,使用yes啟用守護進(jìn)程
daemonize no
# When running daemonized, Redis writes a pid file in /var/run/redis.pid by
# default. You can specify a custom pid file location here.
#當(dāng) Redis 以守護進(jìn)程的方式運行的時候,Redis 默認(rèn)會把 pid 文件放在/var/run/redis.pid
#可配置到其他地址窒所,當(dāng)運行多個 redis 服務(wù)時鹉勒,需要指定不同的 pid 文件和端口
pidfile /var/run/redis.pid
# Accept connections on the specified port, default is 6379.
# If port 0 is specified Redis will not listen on a TCP socket.
#端口
port 6379
# If you want you can bind a single interface, if the bind option is not
# specified all the interfaces will listen for incoming connections.
#指定Redis可接收請求的IP地址,不設(shè)置將處理所有請求,建議生產(chǎn)環(huán)境中設(shè)置
# bind 127.0.0.1
# Close the connection after a client is idle for N seconds (0 to disable)
#客戶端連接的超時時間,單位為秒,超時后會關(guān)閉連接
timeout 0
# Set server verbosity to 'debug'
# it can be one of:
# debug (a lot of information, useful for development/testing)
# verbose (many rarely useful info, but not a mess like the debug level)
# notice (moderately verbose, what you want in production probably)
# warning (only very important / critical messages are logged)
#日志記錄等級,4個可選值
loglevel notice
# Specify the log file name. Also 'stdout' can be used to force
# Redis to log on the standard output. Note that if you use standard
# output for logging but daemonize, logs will be sent to /dev/null
#配置 log 文件地址,默認(rèn)打印在命令行終端的窗口上吵取,也可設(shè)為/dev/null屏蔽日志禽额、
logfile stdout
# Set the number of databases. The default database is DB 0, you can select
# a different one on a per-connection basis using SELECT where
# dbid is a number between 0 and 'databases'-1
#設(shè)置數(shù)據(jù)庫的個數(shù),可以使用 SELECT 命令來切換數(shù)據(jù)庫。
databases 16
#
# Save the DB on disk:
#
# save
#
# Will save the DB if both the given number of seconds and the given
# number of write operations against the DB occurred.
#
# In the example below the behaviour will be to save:
# after 900 sec (15 min) if at least 1 key changed
# after 300 sec (5 min) if at least 10 keys changed
# after 60 sec if at least 10000 keys changed
#
# Note: you can disable saving at all commenting all the "save" lines.
#設(shè)置 Redis 進(jìn)行數(shù)據(jù)庫鏡像的頻率皮官。保存數(shù)據(jù)到disk的策略
#900秒之內(nèi)有1個keys發(fā)生變化時
#30秒之內(nèi)有10個keys發(fā)生變化時
#60秒之內(nèi)有10000個keys發(fā)生變化時
save 900 1
save 300 10
save 60 10000
# Compress string objects using LZF when dump .rdb databases?
# For default that's set to 'yes' as it's almost always a win.
# If you want to save some CPU in the saving child set it to 'no' but
# the dataset will likely be bigger if you have compressible values or keys.
#在進(jìn)行鏡像備份時,是否進(jìn)行壓縮
rdbcompression yes
# The filename where to dump the DB
#鏡像備份文件的文件名
dbfilename dump.rdb
# The working directory.
#
# The DB will be written inside this directory, with the filename specified
# above using the 'dbfilename' configuration directive.
#
# Also the Append Only File will be created inside this directory.
#
# Note that you must specify a directory here, not a file name.
#數(shù)據(jù)庫鏡像備份的文件放置的路徑
#路徑跟文件名分開配置是因為 Redis 備份時脯倒,先會將當(dāng)前數(shù)據(jù)庫的狀態(tài)寫入到一個臨時文件
#等備份完成時,再把該臨時文件替換為上面所指定的文件
#而臨時文件和上面所配置的備份文件都會放在這個指定的路徑當(dāng)中
#默認(rèn)值為 ./
dir /var/lib/redis/
# Master-Slave replication. Use slaveof to make a Redis instance a copy of
# another Redis server. Note that the configuration is local to the slave
# so for example it is possible to configure the slave to save the DB with a
# different interval, or to listen to another port, and so on.
#設(shè)置該數(shù)據(jù)庫為其他數(shù)據(jù)庫的從數(shù)據(jù)庫
#slaveof <masterip> <masterport> 當(dāng)本機為從服務(wù)時捺氢,設(shè)置主服務(wù)的IP及端口
# slaveof
# If the master is password protected (using the "requirepass" configuration
# directive below) it is possible to tell the slave to authenticate before
# starting the replication synchronization process, otherwise the master will
# refuse the slave request.
#指定與主數(shù)據(jù)庫連接時需要的密碼驗證
#masterauth <master-password> 當(dāng)本機為從服務(wù)時藻丢,設(shè)置主服務(wù)的連接密碼
# masterauth
# When a slave lost the connection with the master, or when the replication
# is still in progress, the slave can act in two different ways:
#
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
# still reply to client requests, possibly with out of data data, or the
# data set may just be empty if this is the first synchronization.
#
# 2) if slave-serve-stale data is set to 'no' the slave will reply with
# an error "SYNC with master in progress" to all the kind of commands
# but to INFO and SLAVEOF.
#當(dāng)slave丟失與master的連接時,或slave仍然在于master進(jìn)行數(shù)據(jù)同步時(未與master保持一致)
#slave可有兩種方式來響應(yīng)客戶端請求:
#1)如果 slave-serve-stale-data 設(shè)置成 'yes'(默認(rèn))摄乒,slave仍會響應(yīng)客戶端請求,此時可能會有問題
#2)如果 slave-serve-stale-data 設(shè)置成 'no'悠反,slave會返回"SYNC with master in progress"錯誤信息,但 INFO 和SLAVEOF命令除外馍佑。
slave-serve-stale-data yes
# Require clients to issue AUTH before processing any other
# commands. This might be useful in environments in which you do not trust
# others with access to the host running redis-server.
#
# This should stay commented out for backward compatibility and because most
# people do not need auth (e.g. they run their own servers).
#
# Warning: since Redis is pretty fast an outside user can try up to
# 150k passwords per second against a good box. This means that you should
# use a very strong password otherwise it will be very easy to break.
#設(shè)置客戶端連接后進(jìn)行任何其他指定前需要使用的密碼
#redis速度相當(dāng)快斋否,一個外部用戶在一秒鐘進(jìn)行150K次密碼嘗試,需指定強大的密碼來防止暴力破解
# requirepass foobared
# Set the max number of connected clients at the same time. By default there
# is no limit, and it's up to the number of file descriptors the Redis process
# is able to open. The special value '0' means no limits.
# Once the limit is reached Redis will close all the new connections sending
# an error 'max number of clients reached'.
#限制同時連接的客戶數(shù)量挤茄。
#當(dāng)連接數(shù)超過這個值時如叼,redis 將不再接收其他連接請求,客戶端嘗試連接時將收到 error 信息
# maxclients 128
# Don't use more memory than the specified amount of bytes.
# When the memory limit is reached Redis will try to remove keys
# accordingly to the eviction policy selected (see maxmemmory-policy).
#
# If Redis can't remove keys according to the policy, or if the policy is
# set to 'noeviction', Redis will start to reply with errors to commands
# that would use more memory, like SET, LPUSH, and so on, and will continue
# to reply to read-only commands like GET.
#
# This option is usually useful when using Redis as an LRU cache, or to set
# an hard memory limit for an instance (using the 'noeviction' policy).
#
# WARNING: If you have slaves attached to an instance with maxmemory on,
# the size of the output buffers needed to feed the slaves are subtracted
# from the used memory count, so that network problems / resyncs will
# not trigger a loop where keys are evicted, and in turn the output
# buffer of slaves is full with DELs of keys evicted triggering the deletion
# of more keys, and so forth until the database is completely emptied.
#
# In short... if you have slaves attached it is suggested that you set a lower
# limit for maxmemory so that there is some free RAM on the system for slave
# output buffers (but this is not needed if the policy is 'noeviction').
#設(shè)置redis能夠使用的最大內(nèi)存穷劈。
#達(dá)到最大內(nèi)存設(shè)置后笼恰,Redis會先嘗試清除已到期或即將到期的Key(設(shè)置過expire信息的key)
#在刪除時,按照過期時間進(jìn)行刪除,最早將要被過期的key將最先被刪除
#如果已到期或即將到期的key刪光歇终,仍進(jìn)行set操作社证,那么將返回錯誤
#此時redis將不再接收寫請求,只接收get請求。
#maxmemory的設(shè)置比較適合于把redis當(dāng)作于類似memcached 的緩存來使用
# maxmemory <bytes>
# By default Redis asynchronously dumps the dataset on disk. If you can live
# with the idea that the latest records will be lost if something like a crash
# happens this is the preferred way to run Redis. If instead you care a lot
# about your data and don't want to that a single record can get lost you should
# enable the append only mode: when this mode is enabled Redis will append
# every write operation received in the file appendonly.aof. This file will
# be read on startup in order to rebuild the full dataset in memory.
#
# Note that you can have both the async dumps and the append only file if you
# like (you have to comment the "save" statements above to disable the dumps).
# Still if append only mode is enabled Redis will load the data from the
# log file at startup ignoring the dump.rdb file.
#
# IMPORTANT: Check the BGREWRITEAOF to check how to rewrite the append
# log file in background when it gets too big.
#redis 默認(rèn)每次更新操作后會在后臺異步的把數(shù)據(jù)庫鏡像備份到磁盤评凝,但該備份非常耗時追葡,且備份不宜太頻繁
#redis 同步數(shù)據(jù)文件是按上面save條件來同步的
#如果發(fā)生諸如拉閘限電、拔插頭等狀況,那么將造成比較大范圍的數(shù)據(jù)丟失
#所以redis提供了另外一種更加高效的數(shù)據(jù)庫備份及災(zāi)難恢復(fù)方式
#開啟append only 模式后,redis 將每一次寫操作請求都追加到appendonly.aof 文件中
#redis重新啟動時,會從該文件恢復(fù)出之前的狀態(tài)奕短。
#但可能會造成 appendonly.aof 文件過大宜肉,所以redis支持BGREWRITEAOF 指令,對appendonly.aof重新整理
appendonly no
# The name of the append only file (default: "appendonly.aof")
##更新日志文件名翎碑,默認(rèn)值為appendonly.aof
# appendfilename appendonly.aof
# The fsync() call tells the Operating System to actually write data on disk
# instead to wait for more data in the output buffer. Some OS will really flush
# data on disk, some other OS will just try to do it ASAP.
#
# Redis supports three different modes:
#
# no: don't fsync, just let the OS flush the data when it wants. Faster.
# always: fsync after every write to the append only log . Slow, Safest.
# everysec: fsync only if one second passed since the last fsync. Compromise.
#
# The default is "everysec" that's usually the right compromise between
# speed and data safety. It's up to you to understand if you can relax this to
# "no" that will will let the operating system flush the output buffer when
# it wants, for better performances (but if you can live with the idea of
# some data loss consider the default persistence mode that's snapshotting),
# or on the contrary, use "always" that's very slow but a bit safer than
# everysec.
#
# If unsure, use "everysec".
#設(shè)置對 appendonly.aof 文件進(jìn)行同步的頻率
#always 表示每次有寫操作都進(jìn)行同步,everysec 表示對寫操作進(jìn)行累積,每秒同步一次谬返。
#no表示等操作系統(tǒng)進(jìn)行數(shù)據(jù)緩存同步到磁盤,都進(jìn)行同步,everysec 表示對寫操作進(jìn)行累積,每秒同步一次
# appendfsync always
appendfsync everysec
# appendfsync no
# Virtual Memory allows Redis to work with datasets bigger than the actual
# amount of RAM needed to hold the whole dataset in memory.
# In order to do so very used keys are taken in memory while the other keys
# are swapped into a swap file, similarly to what operating systems do
# with memory pages.
#
# To enable VM just set 'vm-enabled' to yes, and set the following three
# VM parameters accordingly to your needs.
#是否開啟虛擬內(nèi)存支持日杈。
#redis 是一個內(nèi)存數(shù)據(jù)庫遣铝,當(dāng)內(nèi)存滿時,無法接收新的寫請求,所以在redis2.0后,提供了虛擬內(nèi)存的支持
#但需要注意的,redis 所有的key都會放在內(nèi)存中莉擒,在內(nèi)存不夠時,只把value 值放入交換區(qū)
#雖使用虛擬內(nèi)存酿炸,但性能基本不受影響,需要注意的是要把vm-max-memory設(shè)置到足夠來放下所有的key
vm-enabled no
# vm-enabled yes
# This is the path of the Redis swap file. As you can guess, swap files
# can't be shared by different Redis instances, so make sure to use a swap
# file for every redis process you are running. Redis will complain if the
# swap file is already in use.
#
# The best kind of storage for the Redis swap file (that's accessed at random)
# is a Solid State Disk (SSD).
#
# *** WARNING *** if you are using a shared hosting the default of putting
# the swap file under /tmp is not secure. Create a dir with access granted
# only to Redis user and configure Redis to create the swap file there.
#設(shè)置虛擬內(nèi)存的交換文件路徑涨冀,不可多個Redis實例共享
vm-swap-file /tmp/redis.swap
# vm-max-memory configures the VM to use at max the specified amount of
# RAM. Everything that deos not fit will be swapped on disk *if* possible, that
# is, if there is still enough contiguous space in the swap file.
#
# With vm-max-memory 0 the system will swap everything it can. Not a good
# default, just specify the max amount of RAM you can in bytes, but it's
# better to leave some margin. For instance specify an amount of RAM
# that's more or less between 60 and 80% of your free RAM.
#設(shè)置開啟虛擬內(nèi)存后,redis將使用的最大物理內(nèi)存大小填硕。
#默認(rèn)為0盛嘿,redis將把他所有能放到交換文件的都放到交換文件中确垫,以盡量少的使用物理內(nèi)存
#即當(dāng)vm-max-memory設(shè)置為0的時候,其實是所有value都存在于磁盤
#在生產(chǎn)環(huán)境下,需要根據(jù)實際情況設(shè)置該值,最好不要使用默認(rèn)的 0
vm-max-memory 0
# Redis swap files is split into pages. An object can be saved using multiple
# contiguous pages, but pages can't be shared between different objects.
# So if your page is too big, small objects swapped out on disk will waste
# a lot of space. If you page is too small, there is less space in the swap
# file (assuming you configured the same number of total swap file pages).
#
# If you use a lot of small objects, use a page size of 64 or 32 bytes.
# If you use a lot of big objects, use a bigger page size.
# If unsure, use the default :)
#設(shè)置虛擬內(nèi)存的頁大小
如果 value 值比較大,如要在 value 中放置博客、新聞之類的所有文章內(nèi)容缺厉,就設(shè)大一點
vm-page-size 32
# Number of total memory pages in the swap file.
# Given that the page table (a bitmap of free/used pages) is taken in memory,
# every 8 pages on disk will consume 1 byte of RAM.
#
# The total swap size is vm-page-size * vm-pages
#
# With the default of 32-bytes memory pages and 134217728 pages Redis will
# use a 4 GB swap file, that will use 16 MB of RAM for the page table.
#
# It's better to use the smallest acceptable value for your application,
# but the default is large in order to work in most conditions.
#設(shè)置交換文件的總的 page 數(shù)量
#注意page table信息是放在物理內(nèi)存中栓辜,每8個page 就會占據(jù)RAM中的 1 個 byte
#總的虛擬內(nèi)存大小 = vm-page-size * vm-pages
vm-pages 134217728
# Max number of VM I/O threads running at the same time.
# This threads are used to read/write data from/to swap file, since they
# also encode and decode objects from disk to memory or the reverse, a bigger
# number of threads can help with big objects even if they can't help with
# I/O itself as the physical device may not be able to couple with many
# reads/writes operations at the same time.
#
# The special value of 0 turn off threaded I/O and enables the blocking
# Virtual Memory implementation.
#設(shè)置 VM IO 同時使用的線程數(shù)量恋拍。
vm-max-threads 4
# Hashes are encoded in a special way (much more memory efficient) when they
# have at max a given numer of elements, and the biggest element does not
# exceed a given threshold. You can configure this limits with the following
# configuration directives.
#redis 2.0后引入了 hash 數(shù)據(jù)結(jié)構(gòu)。
#hash 中包含超過指定元素個數(shù)并且最大的元素當(dāng)沒有超過臨界時藕甩,hash 將以zipmap來存儲
#zipmap又稱為 small hash施敢,可大大減少內(nèi)存的使用
hash-max-zipmap-entries 512
hash-max-zipmap-value 64
# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
# order to help rehashing the main Redis hash table (the one mapping top-level
# keys to values). The hash table implementation redis uses (see dict.c)
# performs a lazy rehashing: the more operation you run into an hash table
# that is rhashing, the more rehashing "steps" are performed, so if the
# server is idle the rehashing is never complete and some more memory is used
# by the hash table.
#
# The default is to use this millisecond 10 times every second in order to
# active rehashing the main dictionaries, freeing memory when possible.
#
# If unsure:
# use "activerehashing no" if you have hard latency requirements and it is
# not a good thing in your environment that Redis can reply form time to time
# to queries with 2 milliseconds delay.
#
# use "activerehashing yes" if you don't have such hard requirements but
# want to free memory asap when possible.
#是否重置Hash表
#設(shè)置成yes后redis將每100毫秒使用1毫秒CPU時間來對redis的hash表重新hash,可降低內(nèi)存的使用
#當(dāng)使用場景有較為嚴(yán)格的實時性需求,不能接受Redis時不時的對請求有2毫秒的延遲的話狭莱,把這項配置為no僵娃。
#如果沒有這么嚴(yán)格的實時性要求,可以設(shè)置為 yes,以便能夠盡可能快的釋放內(nèi)存
activerehashing yes
Redis官方文檔對VM的使用提出了一些建議:
- 當(dāng)key很小而value很大時,使用VM的效果會比較好.因為這樣節(jié)約的內(nèi)存比較大
- 當(dāng)key不小時,可以考慮使用一些非常方法將很大的key變成很大的value,如可將key,value組合成一個新的value
- 最好使用linux ext3 等對稀疏文件支持比較好的文件系統(tǒng)保存你的swap文件
- vm-max-threads參數(shù)可設(shè)置訪問swap文件的線程數(shù),最好不要超過機器的核數(shù)腋妙;設(shè)置為0則所有對swap文件的操作都是串行的默怨,可能會造成比較長時間的延遲,但是對數(shù)據(jù)完整性有很好的保證
redis數(shù)據(jù)存儲
redis的存儲分為內(nèi)存存儲、磁盤存儲和log文件三部分骤素,配置文件中有三個參數(shù)對其進(jìn)行配置匙睹。
- save seconds updates愚屁,save配置,指出在多長時間內(nèi)痕檬,有多少次更新操作霎槐,就將數(shù)據(jù)同步到數(shù)據(jù)文件∶蚊眨可多個條件配合丘跌,默認(rèn)配置了三個條件。
- appendonly yes/no 唁桩,appendonly配置闭树,指出是否在每次更新操作后進(jìn)行日志記錄,如果不開啟荒澡,可能會在斷電時導(dǎo)致一段時間內(nèi)的數(shù)據(jù)丟失报辱。因為redis本身同步數(shù)據(jù)文件是按上面的save條件來同步的,所以有的數(shù)據(jù)會在一段時間內(nèi)只存在于內(nèi)存中仰猖。
- appendfsync no/always/everysec 捏肢,appendfsync配置,no表示等操作系統(tǒng)進(jìn)行數(shù)據(jù)緩存同步到磁盤饥侵,always表示每次更新操作后手動調(diào)用fsync()將數(shù)據(jù)寫到磁盤鸵赫,everysec表示每秒同步一次。