1. 程式人生 > >FastDFS安裝與配置

FastDFS安裝與配置

FastDFS安裝與配置

簡介:FastDFS是一個開源的輕量級分散式檔案系統,它對檔案進行管理,功能包括:檔案儲存、檔案同步、檔案訪問(檔案上傳、檔案下載)等,解決了大容量儲存和負載均衡的問題。特別適合以檔案為載體的線上服務,如相簿網站、視訊網站等等。
FastDFS為網際網路量身定製,充分考慮了冗餘備份、負載均衡、線性擴容等機制,並注重高可用、高效能等指標,使用FastDFS很容易搭建一套高效能的檔案伺服器叢集提供檔案上傳、下載等服務。

我們在兩臺機器上安裝fastDfs.,兩臺機器的主機名配置資訊如下:

編號 裝置名 IP地址 作業系統
1 slaver2 192.168.50.202 centos7
2 slaver3 192.168.50.203 centos7

在跟蹤節點slaver3和儲存節點slaver2分別安裝FastDfs

(1)下載安裝FastDfs需要的包,並變異安裝

$# sudo yum install make cmake gcc gcc-c++
$#  wget http://www.monkey.org/~provos/libevent-1.4.14b-stable.tar.gz 
$# tar -xzvf  libevent-1.4.14b-stable.tar.gz
$# cd libevent-1.4.14b-stable
$# ./configure –prefix=/usr
$# make
$# sudo make install 
#驗證是否安裝成功
$# ls -al /usr/lib | grep libevent

(2)首先下載fastDfs的包FastDFS_v3.06.tar.gz

$# tar -xzvf   FastDFS_v3.06.tar.gz
$# cd FastDFS
  •  

這裡寫圖片描述

$# ./make.sh

$# sudo ./make install 
  •  

這裡寫圖片描述
//此時表明fastDfs已經安裝成功
tracker節點slaver3配置tracker.conf檔案

#配置tracker server(Tracker Server配置)

#進入目錄 
 $# cd /etc/fdfs/
 $# 建立fastdfs的儲存資訊等
$# sudo mkdir -p /home/sfs/fastdfs_tracker 
$# sudo mkdir -p /home/sfs/fastdfs_storage 
#編輯tracker.conf檔案
sudo vi tracker.conf
#更改配置
base_path=/home/sfs/fastdfs_tracker (自己手動建立的目錄,用來儲存統計儲存的檔案資訊的相關資料和對應的日誌檔案)

1.i
2.返回鍵刪除,貼上
3. esc
4. wq!
  •  

以下為修改後的tracker.conf檔案

# is this config file disabled
# false for enabled
# true for disabled
disabled=false

# bind an address of this host
# empty for bind all addresses of this host
bind_addr=

# the tracker server port
port=22122

# connect timeout in seconds
# default value is 30s
connect_timeout=30

# network timeout in seconds
# default value is 30s
network_timeout=60

# the base path to store data and log files
#base_path=/home/yuqing/fastdfs
base_path=/home/sfs/fastdfs_tracker

# max concurrent connections this server supported
max_connections=256

# work thread count, should <= max_connections
# default value is 4
# since V2.00
work_threads=4

# the method of selecting group to upload files
# 0: round robin
# 1: specify group
# 2: load balance, select the max free space group to upload file
store_lookup=0

# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
store_group=group2

# which storage server to upload file
# 0: round robin (default)
# 1: the first server order by ip address
# 2: the first server order by priority (the minimal)
store_server=0

# which path(means disk or mount point) of the storage server to upload file
# 0: round robin
# 2: load balance, select the max free space path to upload file
store_path=0

# which storage server to download file
# 0: round robin (default)
# 1: the source storage server which the current file uploaded to
download_server=0

# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in 
# a group <= reserved_storage_space, 
# no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
reserved_storage_space = 4GB

#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info

#unix group name to run this program, 
#not set (empty) means run by the group of current user
run_by_group=

#unix username to run this program,
#not set (empty) means run by current user
run_by_user=

# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
allow_hosts=*

# sync log buff to disk every interval seconds
# default value is 10 seconds
sync_log_buff_interval = 10

# check storage server alive interval seconds
check_active_interval = 120

# thread stack size, should >= 64KB
# default value is 64KB
thread_stack_size = 64KB

# auto adjust when the ip address of the storage server changed
# default value is true
storage_ip_changed_auto_adjust = true

# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
storage_sync_file_max_delay = 86400

# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
storage_sync_file_max_time = 300

# if use a trunk file to store several small files
# default value is false
# since V3.00
use_trunk_file = false 

# the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
slot_min_size = 256

# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <=  this value
# default value is 16MB
# since V3.00
slot_max_size = 16MB

# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
trunk_file_size = 64MB

# if create trunk file advancely
# default value is false
trunk_create_file_advance = false

# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
trunk_create_file_time_base = 02:00

# the interval of create trunk file, unit: second
# default value is 38400 (one day)
trunk_create_file_interval = 86400

# the threshold to create trunk file
# when the free trunk file size less than the threshold, will create 
# the trunk files
# default value is 0
trunk_create_file_space_threshold = 20G


# HTTP settings
http.disabled=false

# HTTP port on this tracker server
http.server_port=8080

# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http.check_alive_interval=30

# check storage HTTP server alive type, values are:
#   tcp : connect to the storge server with HTTP port only, 
#        do not request and get response
#   http: storage check alive url must return http status 200
# default value is tcp
http.check_alive_type=tcp

# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http.check_alive_uri=/status.html

# if need find content type from file extension name
http.need_find_content_type=true

#use "#include" directive to include http other settings
##include http.conf

  •  
  •  

(3)store節點slaver2配置storeage.conf檔案,修改它/etc/fdfs目錄下的storage.conf
#配置storage server(Storage Server配置)

$# mkdir -p /home/sfs/fastdfs_tracker 
$# mkdir -p /home/sfs/fastdfs_storage 

#進入目錄
$#  groupadd group1   #建立使用者組
$# cd /etc/fdfs

#編輯storage.conf檔案

$# vi storage.conf

#更改配置
group_name=group1 (用於存檔案的組名,自己命名)

base_path=/home/sfs/fastdfs_storage 
(自己手動建立的目錄,用來儲存上傳或下載的檔案和對應的日誌檔案)

store_path0=/home/sfs/fastdfs_storage (配置與上面一致即可)

# tracker_server 多個則配置多行

tracker_server=slaver3:22122 (安裝tracker_server的機器ip及tracker.conf檔案中配置的埠,預設是22122)
  •  

以下為修改後的storage.conf檔案

# is this config file disabled
# false for enabled
# true for disabled
disabled=false

# the name of the group this storage server belongs to
group_name=group1

# bind an address of this host
# empty for bind all addresses of this host
bind_addr=

# if bind an address of this host when connect to other servers 
# (this storage server as a client)
# true for binding the address configed by above parameter: "bind_addr"
# false for binding any address of this host
client_bind=true

# the storage server port
port=23000

# connect timeout in seconds
# default value is 30s
connect_timeout=30

# network timeout in seconds
# default value is 30s
network_timeout=60

# heart beat interval in seconds
heart_beat_interval=30

# disk usage report interval in seconds
stat_report_interval=60

# the base path to store data and log files
#base_path=/home/yuqing/fastdfs
base_path=/home/sfs/fastdfs_storage 
# max concurrent connections the server supported
# default value is 256
# more max_connections means more memory will be used
max_connections=256

# the buff size to recv / send data
# this parameter must more than 8KB
# default value is 64KB
# since V2.00
buff_size = 256KB

# work thread count, should <= max_connections
# work thread deal network io
# default value is 4
# since V2.00
work_threads=4

# if disk read / write separated
##  false for mixed read and write
##  true for separated read and write
# default value is true
# since V2.00
disk_rw_separated = true

# if read / write file directly
# if set to true, open file will add the O_DIRECT flag to avoid file caching
# by the file system. be careful to set this parameter.
# default value is false
disk_rw_direct = false

# disk reader thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_reader_threads = 1

# disk writer thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_writer_threads = 1

# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
sync_wait_msec=50

# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
sync_interval=0

# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time=00:00

# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_end_time=23:59

# write to the mark file after sync N files
# default value is 500
write_mark_file_freq=500

# path(disk or mount point) count, default value is 1
store_path_count=1

# store_path#, based 0, if store_path0 not exists, it's value is base_path
# the paths must be exist
store_path0=/home/sfs/fastdfs_storage 
#store_path1=/home/yuqing/fastdfs2

# subdir_count  * subdir_count directories will be auto created under each 
# store_path (disk), value can be 1 to 256, default value is 256
subdir_count_per_path=256

# tracker_server can ocur more than once, and tracker_server format is
#  "host:port", host can be hostname or ip address
tracker_server=localhost:22122

#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info

#unix group name to run this program, 
#not set (empty) means run by the group of current user
run_by_group=

#unix username to run this program,
#not set (empty) means run by current user
run_by_user=

# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
allow_hosts=*

# the mode of the files distributed to the data path
# 0: round robin(default)
# 1: random, distributted by hash code
file_distribute_path_mode=0

# valid when file_distribute_to_path is set to 0 (round robin), 
# when the written file count reaches this number, then rotate to next path
# default value is 100
file_distribute_rotate_count=100

# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
fsync_after_written_bytes=0

# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
sync_log_buff_interval=10

# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
sync_binlog_buff_interval=10

# sync storage stat info to disk every interval seconds
# default value is 300 seconds
sync_stat_file_interval=300

# thread stack size, should >= 512KB
# default value is 512KB
thread_stack_size=512KB

# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
upload_priority=10

# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix=

# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
check_file_duplicate=0

# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
key_namespace=FastDFS

# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
keep_alive=0

# you can use "#include filename" (not include double quotes) directive to 
# load FastDHT server list, when the filename is a relative path such as 
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf


#HTTP settings
http.disabled=false

# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
http.domain_name=

# the port of the web server on this storage server
http.server_port=8888

http.trunk_size=256KB

# if need find content type from file extension name
http.need_find_content_type=true

#use "#include" directive to include HTTP other settings
##include http.conf

  •  
  •  
  •  
  •  

(4)啟動

tracker節點slaver3的開啟


tracker server 和storage server,需要注意的是先啟動tracker server。

#啟動tracker server

sudo /usr/local/bin/fdfs_trackerd  /etc/fdfs/tracker.conf
此時如果報錯如下

這裡寫圖片描述
則建立軟連線解決:

$#   ln -s /usr/lib/libevent-1.4.so.2 /usr/lib64/libevent-1.4.so.2
# 再次開啟
$# sudo /usr/local/bin/fdfs_trackerd /etc/fdfs/tracker.conf 

#檢驗是否開啟成功
$# netstat -nplt |grep 22122 

這裡寫圖片描述

此時表明tracker節點已經開啟成功了


# 儲存節點節點slaver2的開啟
#啟動storage server
$# sudo /usr/local/bin/fdfs_storaged  /etc/fdfs/storage.conf
$# netstat -nplt |grep 23000

這裡寫圖片描述

此時表明儲存節點已經安裝成功了

下面我們檢視叢集的資訊

#測試安裝檢視資訊
$# /usr/local/bin/fdfs_monitor /etc/fdfs/storage.conf

這裡寫圖片描述
此時表明,fastDsfs的叢集已經搭建完成。

(5)驗證工作
現在我們來上傳一張照片在測試,以驗證整個叢集已經搭建成功
我們在跟蹤節點slaver3的root下穿件一個檔案,內容為“helloo FastDfs”

$# vi test.txt
hello FastDfs

修改/etc/fdfs/client.conf
$#vi /etc/fdfs/client.conf
base_path=/home/sfs/fastdfs_tracker //御前面的跟蹤節點的路徑保持一直
tracker_server=slaver2:22122

#上傳
$# /usr/local/bin/fdfs_upload_file  /etc/fdfs/client.conf  /root/test.txt

這裡寫圖片描述
此時以上傳完成

我們來到儲存節點slaver2檢視該檔案是否可以訪問到

cd /home/sfs/fastdfs_storage/data/00/00
vi wKgyylpTeonpM8s2AAAAC7UzAkk485.txt
  • 1
  • 2

這裡寫圖片描述

這裡寫圖片描述

我們發現他確實已經上傳完成。至此整個FastDfs叢集已經安裝完畢,後續增加儲存節點或者跟蹤節點,只需要使他們的配置檔案保持一致即可,非常方便快捷。