Java 開發(fā)人員常用的服務(wù)配置(Nginx洲赵、Tomcat土铺、JVM、Mysql板鬓、Redis)

Nginx

Nginx是一款由C語言編寫的高性能悲敷、輕量級的HTTP和反向代理服務(wù)器,同時也是一款I(lǐng)MAP/POP3/SMTP服務(wù)器俭令。

  1. nginx.conf:Nginx核心配置文件后德,linux下默認(rèn)安裝在/etc/nginx/
# Nginx所用用戶和組,window下不指定  
user  www-data;  
# 工作的子進程數(shù)量(通常等于CPU數(shù)量或者2倍于CPU)  
worker_processes  auto;  
# pid存放文件  
pid  /run/nginx.pid; 

# 簡化調(diào)試 此指令不得用于生產(chǎn)環(huán)境 
#master_process  off;
# 簡化調(diào)試 此指令可以用到生產(chǎn)環(huán)境
#daemon off;

# 最大文件描述符
worker_rlimit_nofile 51200;

events {  
    # 使用網(wǎng)絡(luò)IO模型linux建議epoll抄腔,F(xiàn)reeBSD建議采用kqueue瓢湃,window下不指定。  
    #use epoll;  

    # 允許最大連接數(shù)  
    worker_connections 1024;  

    # 此指令的作用是立即接受所有連接放到監(jiān)聽隊列中赫蛇,使得Nginx Worker能夠在獲得新連接通知時盡可能多的接受連接
    #multi_accept on;
} 

# load modules compiled as Dynamic Shared Object (DSO)
#
#dso {
#    load ngx_http_fastcgi_module.so;
#    load ngx_http_rewrite_module.so;
#}

http {  

    # 反向代理相關(guān)配置(自行新建)绵患,參考下面的proxy.conf
    #include  /etc/nginx/conf.d/proxy.conf;;
    # GZIP壓縮相關(guān)配置(自行新建),參考下面的gzip.conf
    #include  /etc/nginx/conf.d//gzip.conf;

    ##
    # Basic Settings
    ##

    sendfile on;
    # 在一個數(shù)據(jù)包里發(fā)送所有頭文件悟耘,而不是一個接一個的發(fā)送
    tcp_nopush on;
    # 不緩存數(shù)據(jù)落蝙,而是一段一段的發(fā)送
    tcp_nodelay on;
    # 設(shè)置客戶端keep-alive超時時間
    keepalive_timeout 65;
    types_hash_max_size 2048;

    client_header_timeout  10;  
    client_body_timeout  10;  
    send_timeout  10;  

    client_header_buffer_size    1k;  
    large_client_header_buffers  4  4k; 

    # 允許客戶端請求的最大單文件字節(jié)數(shù)
    client_max_body_size 10m; 

    client_body_in_single_buffer on;
    # 緩沖區(qū)代理緩沖用戶端請求的最大字節(jié)數(shù)
    client_body_buffer_size 128k;

    # 關(guān)閉Nginx的版本號和系統(tǒng)發(fā)行版本顯示,默認(rèn)是on
    # server_tokens off;

    # server_names_hash_bucket_size 64;
    # server_name_in_redirect off;

    # 設(shè)置文件使用的默認(rèn)的MIME-type
    include  /etc/nginx/mime.types;
    default_type  application/octet-stream;  

    ##
    # SSL Settings
    ##

    ssl_protocols TLSv1 TLSv1.1 TLSv1.2; # Dropping SSLv3, ref: POODLE
    ssl_prefer_server_ciphers on;

    ##
    # Logging Settings
    ##

    # 自定義access_log日志格式和級別  
    #log_format  main  '$remote_addr - $remote_user [$time_local] $request "$status" $body_bytes_sent "$http_referer" "$http_user_agent" "$http_x_forwarded_for"'; 

    # 關(guān)閉日志可配置off
    # 訪問和錯誤日志存放路徑暂幼,常見日志級別有[ debug | info | notice | warn | error | crit | alert | emerg ]筏勒,級別越高記錄的信息越少。 
    access_log /var/log/nginx/access.log main;
    error_log /var/log/nginx/error.log notice;

    # 負(fù)載均衡配置
    upstream upstream_test{  

        # 負(fù)載策略旺嬉,常見有輪詢(默認(rèn))管行、指定權(quán)重(weight 默認(rèn)為1.weight越大,負(fù)載的權(quán)重就越大)邪媳、IP綁定(ip_hash)捐顷、fair(第三方)荡陷、url_hash(第三方)
        #ip_hash;

        server 127.0.0.1:9090 down; (down 表示當(dāng)前server暫時不參與負(fù)載) 
        server 127.0.0.1:8080 weight=10;
        server 127.0.0.1:8081 weight=5;
        server 127.0.0.1:7070 backup; (其它所有的非backup機器down掉或者忙碌時候,請求backup機器) 

        ## Tengine config
        #check interval=300 rise=10 fall=10 timeout=100 type=http port=80;
        #check_http_send "GET / HTTP/1.0\r\n\r\n";
        #check_http_expect_alive http_2xx http_3xx;

        ## Tengine config
        #session_sticky cookie=cookieTest mode=insert;
    }  

    # server配置迅涮,同時支持監(jiān)聽80和443端口
    server {  
        listen  80;
        # https配置
        listen 443 ssl http2; #default_server;
        listen [::]:443 ssl http2; #default_server;
        ssl_certificate "/etc/nginx/ssh_cert/1_domain.cn_bundle.crt";
        ssl_certificate_key "/etc/nginx/ssh_cert/2_domain.cn.key";
        ssl_session_cache shared:SSL:1m;
        ssl_session_timeout  10m;
        ssl_ciphers HIGH:!aNULL:!MD5;
        ssl_prefer_server_ciphers on;

        # 可使用通配符形式配置亲善,如*.domain.com;
        server_name  domain.com  www.domain.com;;  

        # 如果訪問時沒有加www,則跳轉(zhuǎn)至www.domain.com
        if ($host !~* www.domain.com) { 
            rewrite ^(.*)$ http://www.domain.com/$1 permanent;
        }

        location / {
            # 指定目錄位置
             root /etc/nginx/; 
            # 開啟目錄瀏覽
            autoindex  on;
            # 默認(rèn)為on逗柴,顯示出文件的確切大小蛹头,單位是bytes。改為off后戏溺,顯示出文件的大概大小渣蜗,單位是kB或者MB或者GB
            autoindex_exact_size  off;
            # 默認(rèn)為off,顯示的文件時間為GMT時間旷祸。改為on后耕拷,顯示的文件時間為文件的服務(wù)器時間
            autoindex_localtime on; 
            # 10m之后下載速度為10k
            #limit_rate_after 10m;
            #limit_rate 10k;
        } 

    # 根據(jù)User-Agent過濾網(wǎng)絡(luò)爬蟲
    location /spider {
            if ($http_user_agent ~* "python|curl|java|wget|httpclient|okhttp") {
                return 503;
            }
    }

        location /proxy {
            proxy_pass http://upstream_test;  
            # 配置跨域訪問
            add_header 'Access-Control-Allow-Headers' 'Content-Type';  
            add_header 'Access-Control-Allow-Origin' '*';  
            add_header 'Access-Control-Allow-Methods' 'GET';
            add_header 'Access-Control-Expose-Headers' 'Access-Control-Allow-Origin,Access-Control-Allow-Credentials';
        } 

        # static resources config
        location ~* ^/static/.*\.(jpg|jpeg|gif|png|html|htm|swf|js|css)$ {
            root /etc/nginx/static/;
            access_log  off;
            expires  30d;
        }

        # 當(dāng)頁面發(fā)生異常時,可以指定跳轉(zhuǎn)到location中托享,也可以跳轉(zhuǎn)到指定URL
        error_page 404 http://www.error404.com/;
        error_page 500 502 503 504  /50x.html;
        location = /50x.html {
            root  html;
        }
    }

}

  1. gzip.conf
        ##
        # Gzip Settings
        ##

        gzip on;
        # IE6或更低版本禁用壓縮
        gzip_disable "msie6";
        # 對數(shù)據(jù)啟用壓縮的最少字節(jié)數(shù)骚烧,建議最少大于1000字節(jié) 
        gzip_min_length 1k;
        gzip_vary on;
        # 允許或者禁止壓縮基于請求和響應(yīng)的響應(yīng)流,設(shè)置為any意味著將會壓縮所有的請求
        gzip_proxied any;
        # 設(shè)置數(shù)據(jù)壓縮的等級闰围,范圍是1-9之間的任意數(shù)值赃绊,9是最慢但壓縮比最大
        gzip_comp_level 6;
        gzip_buffers 16 8k;
        gzip_http_version 1.1;
        #指定壓縮的文件類型 
        gzip_types application/font-woff text/plain application/javascript application/json text/css application/xml text/javascript image/jpg image/jpeg image/png image/gif image/x-icon;

  1. proxy.conf
#后端的Web服務(wù)器可以通過X-Forwarded-For獲取用戶真實IP
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;

#nginx跟后端服務(wù)器連接超時時間(代理連接超時)
proxy_connect_timeout  90;

#連接成功后,后端服務(wù)器響應(yīng)時間(代理接收超時)
proxy_read_timeout  90;

#nginx發(fā)送數(shù)據(jù)到后端服務(wù)器超時時間(代理發(fā)送超時)
proxy_send_timeout  600; 

#設(shè)置代理服務(wù)器(nginx)保存用戶頭信息的緩沖區(qū)大小
proxy_buffer_size 4k;

#proxy_buffers緩沖區(qū)羡榴,網(wǎng)頁平均在32k以下的話碧查,這樣設(shè)置
proxy_buffers 4 32k;

#高負(fù)荷下緩沖大小(proxy_buffers*2)
proxy_busy_buffers_size 64k; 

#設(shè)定緩存文件夾大小校仑,大于這個值忠售,將從upstream服務(wù)器傳
proxy_temp_file_write_size 64k;

#為了支持新的upstream keepalive選項
proxy_http_version 1.1;
proxy_set_header Connection "";

Tomcat

Java EE開發(fā)中常用的一款Web應(yīng)用服務(wù)器,也是一個免費開源的Servlet容器迄沫,類似的還有Jetty稻扬、Undertow、Netty等羊瘩。

server.xml

<Connector port="8080" protocol="org.apache.coyote.http11.Http11AprProtocol"
               connectionTimeout="5000"  
               maxThreads="500" 
               minSpareThreads="20"
               acceptCount="50"
               maxConnections="8192"
               redirectPort="8443" />

JVM

關(guān)于JVM參數(shù)配置和調(diào)優(yōu)可以參考: 一只懂JVM參數(shù)的狐貍

Redis

Redis是一款使用C語言編寫的高性能Key-Value開源數(shù)據(jù)庫泰佳,支持存儲的value類型包括有string(字符串)、list(鏈表)困后、set(集合)乐纸、zset(sorted set , 有序集合)和hash(哈希類型)衬廷。

redis.conf: redis核心配置文件

# By default Redis does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
# 設(shè)置此選項使得redis以守護進程方式運行
daemonize yes

# When running daemonized, Redis writes a pid file in /var/run/redis.pid by
# default. You can specify a custom pid file location here.
# 以守護進程運行時摇予,pid的存放路徑
pidfile /var/run/redis.pid

# Accept connections on the specified port, default is 6379.
# If port 0 is specified Redis will not listen on a TCP socket.
# 端口
port 6379

# If you want you can bind a single interface, if the bind option is not
# specified all the interfaces will listen for incoming connections.
# 指定Redis可接收請求的IP地址,不設(shè)置將處理所有請求,建議生產(chǎn)環(huán)境中設(shè)置
# bind 127.0.0.1

# Close the connection after a client is idle for N seconds (0 to disable)
# 客戶端連接的超時時間,單位為秒,超時后會關(guān)閉連接
timeout 0

# Specify the log file name. Also 'stdout' can be used to force
# Redis to log on the standard output. Note that if you use standard
# output for logging but daemonize, logs will be sent to /dev/null
# 配置 log 文件地址,默認(rèn)打印在命令行終端的窗口上
logfile stdout

# Set the number of databases. The default database is DB 0, you can select
# a different one on a per-connection basis using SELECT <dbid> where
# dbid is a number between 0 and 'databases'-1
# 設(shè)置數(shù)據(jù)庫的個數(shù),可以使用 SELECT <dbid>命令來切換數(shù)據(jù)庫。默認(rèn)使用的數(shù)據(jù)庫是 0
databases 16

#
# Save the DB on disk:
#
# save <seconds> <changes>
#
# Will save the DB if both the given number of seconds and the given
# number of write operations against the DB occurred.
#
# In the example below the behaviour will be to save:
# after 900 sec (15 min) if at least 1 key changed
# after 300 sec (5 min) if at least 10 keys changed
# after 60 sec if at least 10000 keys changed
#
# Note: you can disable saving at all commenting all the "save" lines.
# RDB備份的頻率吗跋。
# 900秒之內(nèi)有1個keys發(fā)生變化時
# 30秒之內(nèi)有10個keys發(fā)生變化時
# 60秒之內(nèi)有10000個keys發(fā)生變化時
save 900 1
save 300 10
save 60 10000

# Compress string objects using LZF when dump .rdb databases?
# For default that's set to 'yes' as it's almost always a win.
# If you want to save some CPU in the saving child set it to 'no' but
# the dataset will likely be bigger if you have compressible values or keys.
# 在進行鏡像備份時,是否進行壓縮
rdbcompression yes

# The filename where to dump the DB
# 鏡像備份文件的文件名
dbfilename dump.rdb

# The working directory.
#
# The DB will be written inside this directory, with the filename specified
# above using the 'dbfilename' configuration directive.
# 
# Also the Append Only File will be created inside this directory.
# 
# Note that you must specify a directory here, not a file name.
# 數(shù)據(jù)庫鏡像備份的文件放置的路徑侧戴。這里的路徑跟文件名要分開配置是因為 Redis 在進行備份時,先會將當(dāng)前數(shù)據(jù)庫的狀態(tài)寫入到一個臨時文件中,等備份完成時,再把該該臨時文件替換為上面所指定的文件,
# 而這里的臨時文件和上面所配置的備份文件都會放在這個指定的路徑當(dāng)中
dir ./

# Master-Slave replication. Use slaveof to make a Redis instance a copy of
# another Redis server. Note that the configuration is local to the slave
# so for example it is possible to configure the slave to save the DB with a
# different interval, or to listen to another port, and so on.
# 設(shè)置該數(shù)據(jù)庫為其他數(shù)據(jù)庫的從數(shù)據(jù)庫
# slaveof <masterip> <masterport>

# If the master is password protected (using the "requirepass" configuration
# directive below) it is possible to tell the slave to authenticate before
# starting the replication synchronization process, otherwise the master will
# refuse the slave request.
# 指定與主數(shù)據(jù)庫連接時需要的密碼驗證
# masterauth <master-password>

# Require clients to issue AUTH <PASSWORD> before processing any other
# commands. This might be useful in environments in which you do not trust
# others with access to the host running redis-server.
#
# This should stay commented out for backward compatibility and because most
# people do not need auth (e.g. they run their own servers).
# 
# Warning: since Redis is pretty fast an outside user can try up to
# 150k passwords per second against a good box. This means that you should
# use a very strong password otherwise it will be very easy to break.
# 設(shè)置客戶端連接后進行任何其他指定前需要使用的密碼宁昭。警告:redis速度相當(dāng)快,一個外部的用戶可以在一秒鐘進行150K次的密碼嘗試,你需要指定非常非常強大的密碼來防止暴力破解。
requirepass foobared

# Set the max number of connected clients at the same time. By default there
# is no limit, and it's up to the number of file descriptors the Redis process
# is able to open. The special value '0' means no limits.
# Once the limit is reached Redis will close all the new connections sending
# an error 'max number of clients reached'.
# 限制同時連接的客戶數(shù)量酗宋。當(dāng)連接數(shù)超過這個值時,redis 將不再接收其他連接請求,客戶端嘗試連接時將收到 error 信息
# maxclients 128

# Don't use more memory than the specified amount of bytes.
# When the memory limit is reached Redis will try to remove keys
# accordingly to the eviction policy selected (see maxmemmory-policy).
#
# If Redis can't remove keys according to the policy, or if the policy is
# set to 'noeviction', Redis will start to reply with errors to commands
# that would use more memory, like SET, LPUSH, and so on, and will continue
# to reply to read-only commands like GET.
#
# This option is usually useful when using Redis as an LRU cache, or to set
# an hard memory limit for an instance (using the 'noeviction' policy).
# 當(dāng)內(nèi)存達到設(shè)置上限時积仗,內(nèi)存的淘汰策略
# maxmemmory-policy [volatile-lru or volatile-tt or volatile-randowm or allkeys-lru or allkeys-random] 

# WARNING: If you have slaves attached to an instance with maxmemory on,
# the size of the output buffers needed to feed the slaves are subtracted
# from the used memory count, so that network problems / resyncs will
# not trigger a loop where keys are evicted, and in turn the output
# buffer of slaves is full with DELs of keys evicted triggering the deletion
# of more keys, and so forth until the database is completely emptied.
#
# In short… if you have slaves attached it is suggested that you set a lower
# limit for maxmemory so that there is some free RAM on the system for slave
# output buffers (but this is not needed if the policy is 'noeviction').
# 設(shè)置redis能夠使用的最大內(nèi)存。當(dāng)內(nèi)存滿了的時候,如果還接收到set命令,redis將先嘗試剔除設(shè)置過expire信息的key,而不管該key的過期時間還沒有到達蜕猫。
# 在刪除時,將按照過期時間進行刪除,最早將要被過期的key將最先被刪除寂曹。如果帶有expire信息的key都刪光了,那么將返回錯誤。
# 這樣,redis將不再接收寫請求,只接收get請求回右。maxmemory的設(shè)置比較適合于把redis當(dāng)作于類似memcached 的緩存來使用
# maxmemory <bytes>

# By default Redis asynchronously dumps the dataset on disk. If you can live
# with the idea that the latest records will be lost if something like a crash
# happens this is the preferred way to run Redis. If instead you care a lot
# about your data and don't want to that a single record can get lost you should
# enable the append only mode: when this mode is enabled Redis will append
# every write operation received in the file appendonly.aof. This file will
# be read on startup in order to rebuild the full dataset in memory.
#
# Note that you can have both the async dumps and the append only file if you
# like (you have to comment the "save" statements above to disable the dumps).
# Still if append only mode is enabled Redis will load the data from the
# log file at startup ignoring the dump.rdb file.
#
# IMPORTANT: Check the BGREWRITEAOF to check how to rewrite the append
# log file in background when it gets too big.
# 默認(rèn)情況下,redis 會在后臺異步的把數(shù)據(jù)庫鏡像備份到磁盤,但是該備份是非常耗時的,而且備份也不能很頻繁,如果發(fā)生諸如拉閘限電隆圆、拔插頭等狀況,那么將造成比較大范圍的數(shù)據(jù)丟失。
# 所以redis提供了另外一種更加高效的數(shù)據(jù)庫備份及災(zāi)難恢復(fù)方式翔烁。
# 開啟appendonly 模式之后,redis 會把所接收到的每一次寫操作請求都追加到appendonly.aof 文件中,當(dāng)redis重新啟動時,會從該文件恢復(fù)出之前的狀態(tài)渺氧。
# 但是這樣會造成 appendonly.aof 文件過大,所以redis還支持了BGREWRITEAOF 指令,對appendonly.aof進行重新整理
appendonly no

# The fsync() call tells the Operating System to actually write data on disk
# instead to wait for more data in the output buffer. Some OS will really flush 
# data on disk, some other OS will just try to do it ASAP.
#
# Redis supports three different modes:
#
# no: don't fsync, just let the OS flush the data when it wants. Faster.
# always: fsync after every write to the append only log . Slow, Safest.
# everysec: fsync only if one second passed since the last fsync. Compromise.
#
# The default is "everysec" that's usually the right compromise between
# speed and data safety. It's up to you to understand if you can relax this to
# "no" that will will let the operating system flush the output buffer when
# it wants, for better performances (but if you can live with the idea of
# some data loss consider the default persistence mode that's snapshotting),
# or on the contrary, use "always" that's very slow but a bit safer than
# everysec.
#
# If unsure, use "everysec".
# 設(shè)置對 appendonly.aof 文件進行同步的頻率。always 表示每次有寫操作都進行同步,everysec 表示對寫操作進行累積,每秒同步一次蹬屹。
# appendfsync always
appendfsync everysec
# appendfsync no

# Virtual Memory allows Redis to work with datasets bigger than the actual
# amount of RAM needed to hold the whole dataset in memory.
# In order to do so very used keys are taken in memory while the other keys
# are swapped into a swap file, similarly to what operating systems do
# with memory pages.
#
# To enable VM just set 'vm-enabled' to yes, and set the following three
# VM parameters accordingly to your needs.
# 是否開啟虛擬內(nèi)存支持侣背。因為 redis 是一個內(nèi)存數(shù)據(jù)庫,而且當(dāng)內(nèi)存滿的時候,無法接收新的寫請求,所以在redis2.0中,提供了虛擬內(nèi)存的支持。
# 但是需要注意的是,redis中,所有的key都會放在內(nèi)存中,在內(nèi)存不夠時,只會把value 值放入交換區(qū)慨默。
# 這樣保證了雖然使用虛擬內(nèi)存,但性能基本不受影響,同時,你需要注意的是你要把vm-max-memory設(shè)置到足夠來放下你的所有的key
vm-enabled no
# vm-enabled yes

# This is the path of the Redis swap file. As you can guess, swap files
# can't be shared by different Redis instances, so make sure to use a swap
# file for every redis process you are running. Redis will complain if the
# swap file is already in use.
#
# The best kind of storage for the Redis swap file (that's accessed at random) 
# is a Solid State Disk (SSD).
#
# *** WARNING *** if you are using a shared hosting the default of putting
# the swap file under /tmp is not secure. Create a dir with access granted
# only to Redis user and configure Redis to create the swap file there.
# 設(shè)置虛擬內(nèi)存的交換文件路徑
vm-swap-file /tmp/redis.swap

# vm-max-memory configures the VM to use at max the specified amount of
# RAM. Everything that deos not fit will be swapped on disk *if* possible, that
# is, if there is still enough contiguous space in the swap file.
#
# With vm-max-memory 0 the system will swap everything it can. Not a good
# default, just specify the max amount of RAM you can in bytes, but it's
# better to leave some margin. For instance specify an amount of RAM
# that's more or less between 60 and 80% of your free RAM.
# 這里設(shè)置開啟虛擬內(nèi)存之后,redis將使用的最大物理內(nèi)存的大小贩耐。默認(rèn)為0,redis將把他所有的能放到交換文件的都放到交換文件中,以盡量少的使用物理內(nèi)存。
# 在生產(chǎn)環(huán)境下,需要根據(jù)實際情況設(shè)置該值,最好不要使用默認(rèn)的 0
vm-max-memory 0

# Redis swap files is split into pages. An object can be saved using multiple
# contiguous pages, but pages can't be shared between different objects.
# So if your page is too big, small objects swapped out on disk will waste
# a lot of space. If you page is too small, there is less space in the swap
# file (assuming you configured the same number of total swap file pages).
#
# If you use a lot of small objects, use a page size of 64 or 32 bytes.
# If you use a lot of big objects, use a bigger page size.
# If unsure, use the default 
# 設(shè)置虛擬內(nèi)存的頁大小,如果你的 value 值比較大,比如說你要在 value 中放置博客厦取、新聞之類的所有文章內(nèi)容,就設(shè)大一點,如果要放置的都是很小的內(nèi)容,那就設(shè)小一點
vm-page-size 32

# Number of total memory pages in the swap file.
# Given that the page table (a bitmap of free/used pages) is taken in memory,
# every 8 pages on disk will consume 1 byte of RAM.
#
# The total swap size is vm-page-size * vm-pages
#
# With the default of 32-bytes memory pages and 134217728 pages Redis will
# use a 4 GB swap file, that will use 16 MB of RAM for the page table.
#
# It's better to use the smallest acceptable value for your application,
# but the default is large in order to work in most conditions.
# 設(shè)置交換文件的總的 page 數(shù)量,需要注意的是,page table信息會放在物理內(nèi)存中,每8個page 就會占據(jù)RAM中的 1 個 byte憔杨。
# 總的虛擬內(nèi)存大小 = vm-page-size * vm-pages
vm-pages 134217728

# Max number of VM I/O threads running at the same time.
# This threads are used to read/write data from/to swap file, since they
# also encode and decode objects from disk to memory or the reverse, a bigger
# number of threads can help with big objects even if they can't help with
# I/O itself as the physical device may not be able to couple with many
# reads/writes operations at the same time.
#
# The special value of 0 turn off threaded I/O and enables the blocking
# Virtual Memory implementation.
# 設(shè)置 VM IO 同時使用的線程數(shù)量。
vm-max-threads 4

# Hashes are encoded in a special way (much more memory efficient) when they
# have at max a given numer of elements, and the biggest element does not
# exceed a given threshold. You can configure this limits with the following
# configuration directives.
# redis 2.0 中引入了 hash 數(shù)據(jù)結(jié)構(gòu)蒜胖。 
# hash 中包含超過指定元素個數(shù)并且最大的元素當(dāng)沒有超過臨界時,hash 將以zipmap(又稱為 small hash大大減少內(nèi)存使用)來存儲,這里可以設(shè)置這兩個臨界值
hash-max-zipmap-entries 512(hash-max-ziplist-entries for Redis >= 2.6)
hash-max-zipmap-value 64(hash-max-ziplist-value for Redis >= 2.6)

# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
# order to help rehashing the main Redis hash table (the one mapping top-level
# keys to values). The hash table implementation redis uses (see dict.c)
# performs a lazy rehashing: the more operation you run into an hash table
# that is rhashing, the more rehashing "steps" are performed, so if the
# server is idle the rehashing is never complete and some more memory is used
# by the hash table.
# 
# The default is to use this millisecond 10 times every second in order to
# active rehashing the main dictionaries, freeing memory when possible.
#
# If unsure:
# use "activerehashing no" if you have hard latency requirements and it is
# not a good thing in your environment that Redis can reply form time to time
# to queries with 2 milliseconds delay.
#
# use "activerehashing yes" if you don't have such hard requirements but
# want to free memory asap when possible.
# 開啟之后,redis 將在每 100 毫秒時使用 1 毫秒的 CPU 時間來對 redis 的 hash 表進行重新 hash,可以降低內(nèi)存的使用消别。
# 當(dāng)你的使用場景中,有非常嚴(yán)格的實時性需要,不能夠接受 Redis 時不時的對請求有 2 毫秒的延遲的話,把這項配置為 no。
# 如果沒有這么嚴(yán)格的實時性要求,可以設(shè)置為 yes,以便能夠盡可能快的釋放內(nèi)存
activerehashing yes

Mysql

my.cnf: mysql核心配置文件

[client]
port = 3306
socket = /var/lib/mysql/mysql.sock 

[mysqld]
datadir=/var/lib/mysql
socket=/var/lib/mysql/mysql.sock
user = mysql
port = 3306
pid-file = /var/lib/mysql/mysql.pid
#配置此項可以追蹤sql執(zhí)行記錄
log=存放日志的路徑/mysql-sql.log
##以下為開啟主從的必要配置
server-id = 1
binlog-do-db=db_nameA #指定對db_nameA記錄二進制日志  
binlog-ignore-db=db_namB #指定不對db_namB記錄二進制日志
binlog_format = MIXED #binlog格式指定
character-set-server = utf8mb4 #指定數(shù)據(jù)默認(rèn)編碼
log_bin = /var/lib/mysql/log/mysql-bin.log
expire_logs_days = 30
character-set-server = utf8mb4
default-storage-engine = InnoDB
#thread connection
max_connections = 1024
max_connect_errors = 1024
# Try number of CPU's*2 for thread_concurrency
thread_concurrency = 8
thread_cache_size = 256
#*network
skip-name-resolve #跳過dns查詢
max_allowed_packet = 1M
#buffer&amp;cache
table_open_cache = 4096
sort_buffer_size = 256K
join_buffer_size = 256K
#query cache
query_cache_limit = 4M
query_cache_size = 4M
query_cache_type = 1
#temptable
tmp_table_size = 64M
max_heap_table_size = 64M

#自增主鍵配置
auto_increment_offset=1
auto_increment_increment=2

#MyISAM
key_buffer_size = 8M
read_buffer_size = 1M
read_rnd_buffer_size = 256K

#Innodb
innodb_log_file_size = 256M
innodb_log_files_in_group = 2
innodb_status_file = 1

innodb_additional_mem_pool_size = 32M
innodb_buffer_pool_size = 5G
innodb_data_file_path = ibdata1:1G:autoextend
innodb_file_per_table = 1

innodb_additional_mem_pool_size = 32M
innodb_buffer_pool_size = 5G
innodb_data_file_path = ibdata1:1G:autoextend
innodb_file_per_table = 1

innodb_force_recovery = 0
#innodb_table_locks
innodb_thread_concurrency = 8

innodb_flush_log_at_trx_commit = 2

innodb_force_recovery = 0
#innodb_table_locks
innodb_thread_concurrency = 8

innodb_flush_log_at_trx_commit = 2

#slow log
slow_query_log=1
long_query_time=1
slow_query_log_file=/var/lib/mysql/log/slow.log

[mysqld_safe]
#error log
log-error = /var/log/mysqld.log
pid-file = /var/lib/mysql/mysql.pid
open-files-limit = 40960

[mysqldump]
quick
max_allowed_packet = 48M

[mysql]
no-auto-rehash
# Remove the next comment character if you are not familiar with SQL
#safe-updates
default-character-set=utf8

[isamchk]
key_buffer = 128M
sort_buffer_size = 128M
read_buffer = 2M
write_buffer = 2M

[myisamchk]
key_buffer = 128M
sort_buffer_size = 128M
read_buffer = 2M
write_buffer = 2M

原文地址:http://www.reibang.com/p/2fe433702d39

?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請聯(lián)系作者
  • 序言:七十年代末台谢,一起剝皮案震驚了整個濱河市寻狂,隨后出現(xiàn)的幾起案子,更是在濱河造成了極大的恐慌朋沮,老刑警劉巖蛇券,帶你破解...
    沈念sama閱讀 206,311評論 6 481
  • 序言:濱河連續(xù)發(fā)生了三起死亡事件,死亡現(xiàn)場離奇詭異樊拓,居然都是意外死亡纠亚,警方通過查閱死者的電腦和手機,發(fā)現(xiàn)死者居然都...
    沈念sama閱讀 88,339評論 2 382
  • 文/潘曉璐 我一進店門筋夏,熙熙樓的掌柜王于貴愁眉苦臉地迎上來蒂胞,“玉大人,你說我怎么就攤上這事条篷∑妫” “怎么了蛤织?”我有些...
    開封第一講書人閱讀 152,671評論 0 342
  • 文/不壞的土叔 我叫張陵,是天一觀的道長鸿染。 經(jīng)常有香客問我指蚜,道長,這世上最難降的妖魔是什么涨椒? 我笑而不...
    開封第一講書人閱讀 55,252評論 1 279
  • 正文 為了忘掉前任摊鸡,我火速辦了婚禮,結(jié)果婚禮上蚕冬,老公的妹妹穿的比我還像新娘柱宦。我一直安慰自己,他們只是感情好播瞳,可當(dāng)我...
    茶點故事閱讀 64,253評論 5 371
  • 文/花漫 我一把揭開白布掸刊。 她就那樣靜靜地躺著,像睡著了一般赢乓。 火紅的嫁衣襯著肌膚如雪忧侧。 梳的紋絲不亂的頭發(fā)上,一...
    開封第一講書人閱讀 49,031評論 1 285
  • 那天牌芋,我揣著相機與錄音蚓炬,去河邊找鬼。 笑死躺屁,一個胖子當(dāng)著我的面吹牛肯夏,可吹牛的內(nèi)容都是我干的。 我是一名探鬼主播犀暑,決...
    沈念sama閱讀 38,340評論 3 399
  • 文/蒼蘭香墨 我猛地睜開眼驯击,長吁一口氣:“原來是場噩夢啊……” “哼!你這毒婦竟也來了耐亏?” 一聲冷哼從身側(cè)響起徊都,我...
    開封第一講書人閱讀 36,973評論 0 259
  • 序言:老撾萬榮一對情侶失蹤,失蹤者是張志新(化名)和其女友劉穎广辰,沒想到半個月后暇矫,有當(dāng)?shù)厝嗽跇淞掷锇l(fā)現(xiàn)了一具尸體,經(jīng)...
    沈念sama閱讀 43,466評論 1 300
  • 正文 獨居荒郊野嶺守林人離奇死亡择吊,尸身上長有42處帶血的膿包…… 初始之章·張勛 以下內(nèi)容為張勛視角 年9月15日...
    茶點故事閱讀 35,937評論 2 323
  • 正文 我和宋清朗相戀三年李根,在試婚紗的時候發(fā)現(xiàn)自己被綠了。 大學(xué)時的朋友給我發(fā)了我未婚夫和他白月光在一起吃飯的照片几睛。...
    茶點故事閱讀 38,039評論 1 333
  • 序言:一個原本活蹦亂跳的男人離奇死亡房轿,死狀恐怖,靈堂內(nèi)的尸體忽然破棺而出,到底是詐尸還是另有隱情冀续,我是刑警寧澤琼讽,帶...
    沈念sama閱讀 33,701評論 4 323
  • 正文 年R本政府宣布必峰,位于F島的核電站洪唐,受9級特大地震影響,放射性物質(zhì)發(fā)生泄漏吼蚁。R本人自食惡果不足惜凭需,卻給世界環(huán)境...
    茶點故事閱讀 39,254評論 3 307
  • 文/蒙蒙 一、第九天 我趴在偏房一處隱蔽的房頂上張望肝匆。 院中可真熱鬧粒蜈,春花似錦、人聲如沸旗国。這莊子的主人今日做“春日...
    開封第一講書人閱讀 30,259評論 0 19
  • 文/蒼蘭香墨 我抬頭看了看天上的太陽能曾。三九已至度硝,卻和暖如春,著一層夾襖步出監(jiān)牢的瞬間寿冕,已是汗流浹背蕊程。 一陣腳步聲響...
    開封第一講書人閱讀 31,485評論 1 262
  • 我被黑心中介騙來泰國打工, 沒想到剛下飛機就差點兒被人妖公主榨干…… 1. 我叫王不留驼唱,地道東北人藻茂。 一個月前我還...
    沈念sama閱讀 45,497評論 2 354
  • 正文 我出身青樓,卻偏偏與公主長得像玫恳,于是被迫代替她去往敵國和親辨赐。 傳聞我的和親對象是個殘疾皇子,可洞房花燭夜當(dāng)晚...
    茶點故事閱讀 42,786評論 2 345

推薦閱讀更多精彩內(nèi)容