Reids 主從配置隶症,數(shù)據(jù)存儲

配置redis主從IP和port:

master and slave ip port
master 127.0.0.1 6379
slave1 127.0.0.1 6380
slave2 127.0.0.1 6381

下載 redis 源碼包

wget http://download.redis.io/releases/redis-3.2.1.tar.gz

解壓縮 并且 安裝redis


# 解壓文件
tar zxvf redis-3.2.1.tar.gz

# 修改名稱
mv redis-3.2.1 redis-3.2.1.master

# 切換目錄
cd redis-3.2.1.master/src

# 執(zhí)行make和make test
make && make test


# 執(zhí)行成功后會提示:
All tests passed without errors!


# 切換到 上級目錄 
cd ..

# 復(fù)制多個 redis 實例 
cp -r redis-3.2.1.master redis-3.2.1.slave1
cp -r redis-3.2.1.master redis-3.2.1.slave2

修改配置 主服務(wù)器(master)的 配置文件

# 切換目錄
cd redis-3.2.1.master 

# 打開 redis.conf(配置文件)  
vim redis.conf

# 修改以下配置為(根據(jù)字段修改  不是覆蓋全部)  然后保存

    port 6379  # 占用的端口號
    
    pidfile /var/run/redis_6379.pid   # 當(dāng)運行多個 redis 服務(wù)時,需要指定不同的 pid 文件和端口
    
    # slaveof <masterip> <masterport>  #  當(dāng)本機為從服務(wù)時,設(shè)置主服務(wù)的IP及端口    我們這個是主服務(wù)器 不需要設(shè)置
    
    logfile "/data/logs/redis.master.log" # log 文件存儲路徑   首先要測試該路徑是否存在
    
    daemonize yes # Redis默認(rèn)不是以守護進(jìn)程的方式運行,可以通過該配置項修改,使用yes啟用守護進(jìn)程
    
    requirepass password  # 設(shè)置訪問密碼   密碼自行設(shè)置 一定要足夠強壯

# 切換到上級目錄
cd ..

修改配置 從服務(wù)器(slave1, slave2)的 配置文件

注意是兩臺實例都要改哦筷频,改動的流程和配置是一樣的,只不過有些配置的值不一樣柱告。
下面是改了其中一個實例截驮,另一個實例也是這么配置的。


# 切換目錄 
cd redis-3.2.1.slave1  

# 打開 redis.conf(配置文件)
vim redis.conf

# 修改以下配置為(根據(jù)字段修改  不是覆蓋全部)  然后保存
    
    port 6380  # slave1 改成  6380  slave2改成 6381
    
    pidfile /var/run/redis_6380.pid   # slave1 改成  6380  slave2改成 6381
    
    slaveof 127.0.0.1 6379  # 配置 主服務(wù)器的 IP 和 端口
    
    masterauth password   # 配置 訪問主服務(wù)器的的密碼 
    
    logfile "/data/logs/redis.master.log" # slave1 將 master 改成  6380际度, slave2改成 6381
    
    daemonize yes # Redis默認(rèn)不是以守護進(jìn)程的方式運行葵袭,可以通過該配置項修改,使用yes啟用守護進(jìn)程
    
    requirepass password  # 設(shè)置訪問密碼   密碼自行設(shè)置 一定要足夠強壯

依次啟動 redis 實例


# 切換目錄
cd redis-3.2.1.master

# 啟動 主服務(wù)器實例
./src/redis-server redis.conf

# 切換到 slave1 
cd ../redis-3.2.1.slave1

# 啟動 從服務(wù)器(slave1)實例
./src/redis-server redis.conf

# 切換到 slave2 
cd ../redis-3.2.1.slave2

# 啟動 從服務(wù)器(slave1)實例
./src/redis-server redis.conf

測試主從關(guān)系

# 打開 主服務(wù)器 命令行
./redis-3.2.1.master/src/redis-cli 

# 設(shè)置一個值
set test_key test

# 獲取 輸入的值
get test_key

# 退出命令行
ctrl + c

# 進(jìn)入 從服務(wù)器 
./redis-3.2.1.slave1/src/redis-cli -h 127.0.0.1 -p 6380

# 獲取剛才在主服務(wù)器上設(shè)置的值
get test_key


Redis配置文件redis.conf參數(shù)說明

# By default Redis does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
#Redis默認(rèn)不是以守護進(jìn)程的方式運行乖菱,可以通過該配置項修改坡锡,使用yes啟用守護進(jìn)程
daemonize no

# When running daemonized, Redis writes a pid file in /var/run/redis.pid by
# default. You can specify a custom pid file location here.
#當(dāng) Redis 以守護進(jìn)程的方式運行的時候,Redis 默認(rèn)會把 pid 文件放在/var/run/redis.pid
#可配置到其他地址窒所,當(dāng)運行多個 redis 服務(wù)時鹉勒,需要指定不同的 pid 文件和端口
pidfile /var/run/redis.pid

# Accept connections on the specified port, default is 6379.
# If port 0 is specified Redis will not listen on a TCP socket.
#端口
port 6379

# If you want you can bind a single interface, if the bind option is not
# specified all the interfaces will listen for incoming connections.
#指定Redis可接收請求的IP地址,不設(shè)置將處理所有請求,建議生產(chǎn)環(huán)境中設(shè)置
# bind 127.0.0.1

# Close the connection after a client is idle for N seconds (0 to disable)
#客戶端連接的超時時間,單位為秒,超時后會關(guān)閉連接
timeout 0

# Set server verbosity to 'debug'
# it can be one of:
# debug (a lot of information, useful for development/testing)
# verbose (many rarely useful info, but not a mess like the debug level)
# notice (moderately verbose, what you want in production probably)
# warning (only very important / critical messages are logged)
#日志記錄等級,4個可選值
loglevel notice

# Specify the log file name. Also 'stdout' can be used to force
# Redis to log on the standard output. Note that if you use standard
# output for logging but daemonize, logs will be sent to /dev/null
#配置 log 文件地址,默認(rèn)打印在命令行終端的窗口上吵取,也可設(shè)為/dev/null屏蔽日志禽额、
logfile stdout

# Set the number of databases. The default database is DB 0, you can select
# a different one on a per-connection basis using SELECT where
# dbid is a number between 0 and 'databases'-1
#設(shè)置數(shù)據(jù)庫的個數(shù),可以使用 SELECT 命令來切換數(shù)據(jù)庫。
databases 16

#
# Save the DB on disk:
#
#   save
#
#   Will save the DB if both the given number of seconds and the given
#   number of write operations against the DB occurred.
#
#   In the example below the behaviour will be to save:
#   after 900 sec (15 min) if at least 1 key changed
#   after 300 sec (5 min) if at least 10 keys changed
#   after 60 sec if at least 10000 keys changed
#
#   Note: you can disable saving at all commenting all the "save" lines.
#設(shè)置 Redis 進(jìn)行數(shù)據(jù)庫鏡像的頻率皮官。保存數(shù)據(jù)到disk的策略
#900秒之內(nèi)有1個keys發(fā)生變化時
#30秒之內(nèi)有10個keys發(fā)生變化時
#60秒之內(nèi)有10000個keys發(fā)生變化時
save 900 1
save 300 10
save 60 10000

# Compress string objects using LZF when dump .rdb databases?
# For default that's set to 'yes' as it's almost always a win.
# If you want to save some CPU in the saving child set it to 'no' but
# the dataset will likely be bigger if you have compressible values or keys.
#在進(jìn)行鏡像備份時,是否進(jìn)行壓縮
rdbcompression yes

# The filename where to dump the DB
#鏡像備份文件的文件名
dbfilename dump.rdb

# The working directory.
#
# The DB will be written inside this directory, with the filename specified
# above using the 'dbfilename' configuration directive.
# 
# Also the Append Only File will be created inside this directory.
# 
# Note that you must specify a directory here, not a file name.
#數(shù)據(jù)庫鏡像備份的文件放置的路徑
#路徑跟文件名分開配置是因為 Redis 備份時脯倒,先會將當(dāng)前數(shù)據(jù)庫的狀態(tài)寫入到一個臨時文件
#等備份完成時,再把該臨時文件替換為上面所指定的文件
#而臨時文件和上面所配置的備份文件都會放在這個指定的路徑當(dāng)中
#默認(rèn)值為 ./
dir /var/lib/redis/

# Master-Slave replication. Use slaveof to make a Redis instance a copy of
# another Redis server. Note that the configuration is local to the slave
# so for example it is possible to configure the slave to save the DB with a
# different interval, or to listen to another port, and so on.
#設(shè)置該數(shù)據(jù)庫為其他數(shù)據(jù)庫的從數(shù)據(jù)庫
#slaveof <masterip> <masterport> 當(dāng)本機為從服務(wù)時捺氢,設(shè)置主服務(wù)的IP及端口
# slaveof

# If the master is password protected (using the "requirepass" configuration
# directive below) it is possible to tell the slave to authenticate before
# starting the replication synchronization process, otherwise the master will
# refuse the slave request.
#指定與主數(shù)據(jù)庫連接時需要的密碼驗證
#masterauth <master-password> 當(dāng)本機為從服務(wù)時藻丢,設(shè)置主服務(wù)的連接密碼
# masterauth

# When a slave lost the connection with the master, or when the replication
# is still in progress, the slave can act in two different ways:
#
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
#    still reply to client requests, possibly with out of data data, or the
#    data set may just be empty if this is the first synchronization.
#
# 2) if slave-serve-stale data is set to 'no' the slave will reply with
#    an error "SYNC with master in progress" to all the kind of commands
#    but to INFO and SLAVEOF.
#當(dāng)slave丟失與master的連接時,或slave仍然在于master進(jìn)行數(shù)據(jù)同步時(未與master保持一致)
#slave可有兩種方式來響應(yīng)客戶端請求:
#1)如果 slave-serve-stale-data 設(shè)置成 'yes'(默認(rèn))摄乒,slave仍會響應(yīng)客戶端請求,此時可能會有問題
#2)如果 slave-serve-stale-data 設(shè)置成 'no'悠反,slave會返回"SYNC with master in progress"錯誤信息,但 INFO 和SLAVEOF命令除外馍佑。
slave-serve-stale-data yes

# Require clients to issue AUTH before processing any other
# commands.  This might be useful in environments in which you do not trust
# others with access to the host running redis-server.
#
# This should stay commented out for backward compatibility and because most
# people do not need auth (e.g. they run their own servers).
# 
# Warning: since Redis is pretty fast an outside user can try up to
# 150k passwords per second against a good box. This means that you should
# use a very strong password otherwise it will be very easy to break.
#設(shè)置客戶端連接后進(jìn)行任何其他指定前需要使用的密碼
#redis速度相當(dāng)快斋否,一個外部用戶在一秒鐘進(jìn)行150K次密碼嘗試,需指定強大的密碼來防止暴力破解
# requirepass foobared

# Set the max number of connected clients at the same time. By default there
# is no limit, and it's up to the number of file descriptors the Redis process
# is able to open. The special value '0' means no limits.
# Once the limit is reached Redis will close all the new connections sending
# an error 'max number of clients reached'.
#限制同時連接的客戶數(shù)量挤茄。
#當(dāng)連接數(shù)超過這個值時如叼,redis 將不再接收其他連接請求,客戶端嘗試連接時將收到 error 信息
# maxclients 128

# Don't use more memory than the specified amount of bytes.
# When the memory limit is reached Redis will try to remove keys
# accordingly to the eviction policy selected (see maxmemmory-policy).
#
# If Redis can't remove keys according to the policy, or if the policy is
# set to 'noeviction', Redis will start to reply with errors to commands
# that would use more memory, like SET, LPUSH, and so on, and will continue
# to reply to read-only commands like GET.
#
# This option is usually useful when using Redis as an LRU cache, or to set
# an hard memory limit for an instance (using the 'noeviction' policy).
#
# WARNING: If you have slaves attached to an instance with maxmemory on,
# the size of the output buffers needed to feed the slaves are subtracted
# from the used memory count, so that network problems / resyncs will
# not trigger a loop where keys are evicted, and in turn the output
# buffer of slaves is full with DELs of keys evicted triggering the deletion
# of more keys, and so forth until the database is completely emptied.
#
# In short... if you have slaves attached it is suggested that you set a lower
# limit for maxmemory so that there is some free RAM on the system for slave
# output buffers (but this is not needed if the policy is 'noeviction').
#設(shè)置redis能夠使用的最大內(nèi)存穷劈。
#達(dá)到最大內(nèi)存設(shè)置后笼恰,Redis會先嘗試清除已到期或即將到期的Key(設(shè)置過expire信息的key)
#在刪除時,按照過期時間進(jìn)行刪除,最早將要被過期的key將最先被刪除
#如果已到期或即將到期的key刪光歇终,仍進(jìn)行set操作社证,那么將返回錯誤
#此時redis將不再接收寫請求,只接收get請求。
#maxmemory的設(shè)置比較適合于把redis當(dāng)作于類似memcached 的緩存來使用
# maxmemory <bytes>

# By default Redis asynchronously dumps the dataset on disk. If you can live
# with the idea that the latest records will be lost if something like a crash
# happens this is the preferred way to run Redis. If instead you care a lot
# about your data and don't want to that a single record can get lost you should
# enable the append only mode: when this mode is enabled Redis will append
# every write operation received in the file appendonly.aof. This file will
# be read on startup in order to rebuild the full dataset in memory.
#
# Note that you can have both the async dumps and the append only file if you
# like (you have to comment the "save" statements above to disable the dumps).
# Still if append only mode is enabled Redis will load the data from the
# log file at startup ignoring the dump.rdb file.
#
# IMPORTANT: Check the BGREWRITEAOF to check how to rewrite the append
# log file in background when it gets too big.
#redis 默認(rèn)每次更新操作后會在后臺異步的把數(shù)據(jù)庫鏡像備份到磁盤评凝,但該備份非常耗時追葡,且備份不宜太頻繁
#redis 同步數(shù)據(jù)文件是按上面save條件來同步的
#如果發(fā)生諸如拉閘限電、拔插頭等狀況,那么將造成比較大范圍的數(shù)據(jù)丟失
#所以redis提供了另外一種更加高效的數(shù)據(jù)庫備份及災(zāi)難恢復(fù)方式
#開啟append only 模式后,redis 將每一次寫操作請求都追加到appendonly.aof 文件中
#redis重新啟動時,會從該文件恢復(fù)出之前的狀態(tài)奕短。
#但可能會造成 appendonly.aof 文件過大宜肉,所以redis支持BGREWRITEAOF 指令,對appendonly.aof重新整理
appendonly no

# The name of the append only file (default: "appendonly.aof")
##更新日志文件名翎碑,默認(rèn)值為appendonly.aof
# appendfilename appendonly.aof

# The fsync() call tells the Operating System to actually write data on disk
# instead to wait for more data in the output buffer. Some OS will really flush 
# data on disk, some other OS will just try to do it ASAP.
#
# Redis supports three different modes:
#
# no: don't fsync, just let the OS flush the data when it wants. Faster.
# always: fsync after every write to the append only log . Slow, Safest.
# everysec: fsync only if one second passed since the last fsync. Compromise.
#
# The default is "everysec" that's usually the right compromise between
# speed and data safety. It's up to you to understand if you can relax this to
# "no" that will will let the operating system flush the output buffer when
# it wants, for better performances (but if you can live with the idea of
# some data loss consider the default persistence mode that's snapshotting),
# or on the contrary, use "always" that's very slow but a bit safer than
# everysec.
#
# If unsure, use "everysec".
#設(shè)置對 appendonly.aof 文件進(jìn)行同步的頻率
#always 表示每次有寫操作都進(jìn)行同步,everysec 表示對寫操作進(jìn)行累積,每秒同步一次谬返。
#no表示等操作系統(tǒng)進(jìn)行數(shù)據(jù)緩存同步到磁盤,都進(jìn)行同步,everysec 表示對寫操作進(jìn)行累積,每秒同步一次
# appendfsync always
appendfsync everysec
# appendfsync no

# Virtual Memory allows Redis to work with datasets bigger than the actual
# amount of RAM needed to hold the whole dataset in memory.
# In order to do so very used keys are taken in memory while the other keys
# are swapped into a swap file, similarly to what operating systems do
# with memory pages.
#
# To enable VM just set 'vm-enabled' to yes, and set the following three
# VM parameters accordingly to your needs.
#是否開啟虛擬內(nèi)存支持日杈。
#redis 是一個內(nèi)存數(shù)據(jù)庫遣铝,當(dāng)內(nèi)存滿時,無法接收新的寫請求,所以在redis2.0后,提供了虛擬內(nèi)存的支持
#但需要注意的,redis 所有的key都會放在內(nèi)存中莉擒,在內(nèi)存不夠時,只把value 值放入交換區(qū)
#雖使用虛擬內(nèi)存酿炸,但性能基本不受影響,需要注意的是要把vm-max-memory設(shè)置到足夠來放下所有的key
vm-enabled no
# vm-enabled yes

# This is the path of the Redis swap file. As you can guess, swap files
# can't be shared by different Redis instances, so make sure to use a swap
# file for every redis process you are running. Redis will complain if the
# swap file is already in use.
#
# The best kind of storage for the Redis swap file (that's accessed at random) 
# is a Solid State Disk (SSD).
#
# *** WARNING *** if you are using a shared hosting the default of putting
# the swap file under /tmp is not secure. Create a dir with access granted
# only to Redis user and configure Redis to create the swap file there.
#設(shè)置虛擬內(nèi)存的交換文件路徑涨冀,不可多個Redis實例共享
vm-swap-file /tmp/redis.swap

# vm-max-memory configures the VM to use at max the specified amount of
# RAM. Everything that deos not fit will be swapped on disk *if* possible, that
# is, if there is still enough contiguous space in the swap file.
#
# With vm-max-memory 0 the system will swap everything it can. Not a good
# default, just specify the max amount of RAM you can in bytes, but it's
# better to leave some margin. For instance specify an amount of RAM
# that's more or less between 60 and 80% of your free RAM.
#設(shè)置開啟虛擬內(nèi)存后,redis將使用的最大物理內(nèi)存大小填硕。
#默認(rèn)為0盛嘿,redis將把他所有能放到交換文件的都放到交換文件中确垫,以盡量少的使用物理內(nèi)存
#即當(dāng)vm-max-memory設(shè)置為0的時候,其實是所有value都存在于磁盤
#在生產(chǎn)環(huán)境下,需要根據(jù)實際情況設(shè)置該值,最好不要使用默認(rèn)的 0
vm-max-memory 0

# Redis swap files is split into pages. An object can be saved using multiple
# contiguous pages, but pages can't be shared between different objects.
# So if your page is too big, small objects swapped out on disk will waste
# a lot of space. If you page is too small, there is less space in the swap
# file (assuming you configured the same number of total swap file pages).
#
# If you use a lot of small objects, use a page size of 64 or 32 bytes.
# If you use a lot of big objects, use a bigger page size.
# If unsure, use the default :)
#設(shè)置虛擬內(nèi)存的頁大小
如果 value 值比較大,如要在 value 中放置博客、新聞之類的所有文章內(nèi)容缺厉,就設(shè)大一點
vm-page-size 32

# Number of total memory pages in the swap file.
# Given that the page table (a bitmap of free/used pages) is taken in memory,
# every 8 pages on disk will consume 1 byte of RAM.
#
# The total swap size is vm-page-size * vm-pages
#
# With the default of 32-bytes memory pages and 134217728 pages Redis will
# use a 4 GB swap file, that will use 16 MB of RAM for the page table.
#
# It's better to use the smallest acceptable value for your application,
# but the default is large in order to work in most conditions.
#設(shè)置交換文件的總的 page 數(shù)量
#注意page table信息是放在物理內(nèi)存中栓辜,每8個page 就會占據(jù)RAM中的 1 個 byte
#總的虛擬內(nèi)存大小 = vm-page-size * vm-pages
vm-pages 134217728

# Max number of VM I/O threads running at the same time.
# This threads are used to read/write data from/to swap file, since they
# also encode and decode objects from disk to memory or the reverse, a bigger
# number of threads can help with big objects even if they can't help with
# I/O itself as the physical device may not be able to couple with many
# reads/writes operations at the same time.
#
# The special value of 0 turn off threaded I/O and enables the blocking
# Virtual Memory implementation.
#設(shè)置 VM IO 同時使用的線程數(shù)量恋拍。
vm-max-threads 4

# Hashes are encoded in a special way (much more memory efficient) when they
# have at max a given numer of elements, and the biggest element does not
# exceed a given threshold. You can configure this limits with the following
# configuration directives.
#redis 2.0后引入了 hash 數(shù)據(jù)結(jié)構(gòu)。 
#hash 中包含超過指定元素個數(shù)并且最大的元素當(dāng)沒有超過臨界時藕甩,hash 將以zipmap來存儲
#zipmap又稱為 small hash施敢,可大大減少內(nèi)存的使用
hash-max-zipmap-entries 512
hash-max-zipmap-value 64

# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
# order to help rehashing the main Redis hash table (the one mapping top-level
# keys to values). The hash table implementation redis uses (see dict.c)
# performs a lazy rehashing: the more operation you run into an hash table
# that is rhashing, the more rehashing "steps" are performed, so if the
# server is idle the rehashing is never complete and some more memory is used
# by the hash table.
# 
# The default is to use this millisecond 10 times every second in order to
# active rehashing the main dictionaries, freeing memory when possible.
#
# If unsure:
# use "activerehashing no" if you have hard latency requirements and it is
# not a good thing in your environment that Redis can reply form time to time
# to queries with 2 milliseconds delay.
#
# use "activerehashing yes" if you don't have such hard requirements but
# want to free memory asap when possible.
#是否重置Hash表
#設(shè)置成yes后redis將每100毫秒使用1毫秒CPU時間來對redis的hash表重新hash,可降低內(nèi)存的使用
#當(dāng)使用場景有較為嚴(yán)格的實時性需求,不能接受Redis時不時的對請求有2毫秒的延遲的話狭莱,把這項配置為no僵娃。
#如果沒有這么嚴(yán)格的實時性要求,可以設(shè)置為 yes,以便能夠盡可能快的釋放內(nèi)存
activerehashing yes

Redis官方文檔對VM的使用提出了一些建議:

  • 當(dāng)key很小而value很大時,使用VM的效果會比較好.因為這樣節(jié)約的內(nèi)存比較大
  • 當(dāng)key不小時,可以考慮使用一些非常方法將很大的key變成很大的value,如可將key,value組合成一個新的value
  • 最好使用linux ext3 等對稀疏文件支持比較好的文件系統(tǒng)保存你的swap文件
  • vm-max-threads參數(shù)可設(shè)置訪問swap文件的線程數(shù),最好不要超過機器的核數(shù)腋妙;設(shè)置為0則所有對swap文件的操作都是串行的默怨,可能會造成比較長時間的延遲,但是對數(shù)據(jù)完整性有很好的保證

redis數(shù)據(jù)存儲

redis的存儲分為內(nèi)存存儲、磁盤存儲和log文件三部分骤素,配置文件中有三個參數(shù)對其進(jìn)行配置匙睹。

  • save seconds updates愚屁,save配置,指出在多長時間內(nèi)痕檬,有多少次更新操作霎槐,就將數(shù)據(jù)同步到數(shù)據(jù)文件∶蚊眨可多個條件配合丘跌,默認(rèn)配置了三個條件。
  • appendonly yes/no 唁桩,appendonly配置闭树,指出是否在每次更新操作后進(jìn)行日志記錄,如果不開啟荒澡,可能會在斷電時導(dǎo)致一段時間內(nèi)的數(shù)據(jù)丟失报辱。因為redis本身同步數(shù)據(jù)文件是按上面的save條件來同步的,所以有的數(shù)據(jù)會在一段時間內(nèi)只存在于內(nèi)存中仰猖。
  • appendfsync no/always/everysec 捏肢,appendfsync配置,no表示等操作系統(tǒng)進(jìn)行數(shù)據(jù)緩存同步到磁盤饥侵,always表示每次更新操作后手動調(diào)用fsync()將數(shù)據(jù)寫到磁盤鸵赫,everysec表示每秒同步一次。
最后編輯于
?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請聯(lián)系作者
  • 序言:七十年代末躏升,一起剝皮案震驚了整個濱河市辩棒,隨后出現(xiàn)的幾起案子,更是在濱河造成了極大的恐慌膨疏,老刑警劉巖一睁,帶你破解...
    沈念sama閱讀 219,490評論 6 508
  • 序言:濱河連續(xù)發(fā)生了三起死亡事件,死亡現(xiàn)場離奇詭異佃却,居然都是意外死亡者吁,警方通過查閱死者的電腦和手機,發(fā)現(xiàn)死者居然都...
    沈念sama閱讀 93,581評論 3 395
  • 文/潘曉璐 我一進(jìn)店門饲帅,熙熙樓的掌柜王于貴愁眉苦臉地迎上來复凳,“玉大人,你說我怎么就攤上這事灶泵∮耍” “怎么了?”我有些...
    開封第一講書人閱讀 165,830評論 0 356
  • 文/不壞的土叔 我叫張陵赦邻,是天一觀的道長髓棋。 經(jīng)常有香客問我,道長,這世上最難降的妖魔是什么按声? 我笑而不...
    開封第一講書人閱讀 58,957評論 1 295
  • 正文 為了忘掉前任膳犹,我火速辦了婚禮,結(jié)果婚禮上儒喊,老公的妹妹穿的比我還像新娘镣奋。我一直安慰自己币呵,他們只是感情好怀愧,可當(dāng)我...
    茶點故事閱讀 67,974評論 6 393
  • 文/花漫 我一把揭開白布。 她就那樣靜靜地躺著余赢,像睡著了一般芯义。 火紅的嫁衣襯著肌膚如雪。 梳的紋絲不亂的頭發(fā)上妻柒,一...
    開封第一講書人閱讀 51,754評論 1 307
  • 那天扛拨,我揣著相機與錄音,去河邊找鬼举塔。 笑死绑警,一個胖子當(dāng)著我的面吹牛,可吹牛的內(nèi)容都是我干的央渣。 我是一名探鬼主播计盒,決...
    沈念sama閱讀 40,464評論 3 420
  • 文/蒼蘭香墨 我猛地睜開眼,長吁一口氣:“原來是場噩夢啊……” “哼芽丹!你這毒婦竟也來了北启?” 一聲冷哼從身側(cè)響起,我...
    開封第一講書人閱讀 39,357評論 0 276
  • 序言:老撾萬榮一對情侶失蹤拔第,失蹤者是張志新(化名)和其女友劉穎咕村,沒想到半個月后,有當(dāng)?shù)厝嗽跇淞掷锇l(fā)現(xiàn)了一具尸體蚊俺,經(jīng)...
    沈念sama閱讀 45,847評論 1 317
  • 正文 獨居荒郊野嶺守林人離奇死亡懈涛,尸身上長有42處帶血的膿包…… 初始之章·張勛 以下內(nèi)容為張勛視角 年9月15日...
    茶點故事閱讀 37,995評論 3 338
  • 正文 我和宋清朗相戀三年,在試婚紗的時候發(fā)現(xiàn)自己被綠了泳猬。 大學(xué)時的朋友給我發(fā)了我未婚夫和他白月光在一起吃飯的照片批钠。...
    茶點故事閱讀 40,137評論 1 351
  • 序言:一個原本活蹦亂跳的男人離奇死亡,死狀恐怖暂殖,靈堂內(nèi)的尸體忽然破棺而出价匠,到底是詐尸還是另有隱情,我是刑警寧澤呛每,帶...
    沈念sama閱讀 35,819評論 5 346
  • 正文 年R本政府宣布踩窖,位于F島的核電站,受9級特大地震影響晨横,放射性物質(zhì)發(fā)生泄漏洋腮。R本人自食惡果不足惜箫柳,卻給世界環(huán)境...
    茶點故事閱讀 41,482評論 3 331
  • 文/蒙蒙 一、第九天 我趴在偏房一處隱蔽的房頂上張望啥供。 院中可真熱鬧悯恍,春花似錦、人聲如沸伙狐。這莊子的主人今日做“春日...
    開封第一講書人閱讀 32,023評論 0 22
  • 文/蒼蘭香墨 我抬頭看了看天上的太陽贷屎。三九已至罢防,卻和暖如春,著一層夾襖步出監(jiān)牢的瞬間唉侄,已是汗流浹背咒吐。 一陣腳步聲響...
    開封第一講書人閱讀 33,149評論 1 272
  • 我被黑心中介騙來泰國打工, 沒想到剛下飛機就差點兒被人妖公主榨干…… 1. 我叫王不留属划,地道東北人恬叹。 一個月前我還...
    沈念sama閱讀 48,409評論 3 373
  • 正文 我出身青樓,卻偏偏與公主長得像同眯,于是被迫代替她去往敵國和親绽昼。 傳聞我的和親對象是個殘疾皇子,可洞房花燭夜當(dāng)晚...
    茶點故事閱讀 45,086評論 2 355

推薦閱讀更多精彩內(nèi)容