Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
Q
qg-dockerfiles
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
QA
qg-dockerfiles
Commits
11371cdb
Commit
11371cdb
authored
Apr 24, 2018
by
kalvin
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
fastdfs
parent
0baf5915
Changes
8
Expand all
Hide whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
1831 additions
and
0 deletions
+1831
-0
Dockerfile
baseimg/tools/fastdfs/Dockerfile
+42
-0
client.conf
baseimg/tools/fastdfs/conf/client.conf
+62
-0
http.conf
baseimg/tools/fastdfs/conf/http.conf
+26
-0
mime.types
baseimg/tools/fastdfs/conf/mime.types
+1065
-0
storage.conf
baseimg/tools/fastdfs/conf/storage.conf
+285
-0
storage_ids.conf
baseimg/tools/fastdfs/conf/storage_ids.conf
+3
-0
tracker.conf
baseimg/tools/fastdfs/conf/tracker.conf
+276
-0
start.sh
baseimg/tools/fastdfs/start.sh
+72
-0
No files found.
baseimg/tools/fastdfs/Dockerfile
0 → 100644
View file @
11371cdb
FROM
192.168.4.4/baseimg/centos:201803291115
ENV
FASTDFS_PATH=/opt/fdfs \
FASTDFS_BASE_PATH=/var/fdfs \
PORT= \
GROUP_NAME= \
TRACKER_SERVER=
#create the dirs to store the files downloaded from internet
RUN
mkdir
-p
${
FASTDFS_PATH
}
/libfastcommon
\
&&
mkdir
-p
${
FASTDFS_PATH
}
/fastdfs
\
&&
mkdir
${
FASTDFS_BASE_PATH
}
#compile the libfastcommon
WORKDIR
${FASTDFS_PATH}/libfastcommon
RUN
git clone
--branch
V1.0.36
--depth
1 https://github.com/happyfish100/libfastcommon.git
${
FASTDFS_PATH
}
/libfastcommon
\
&&
./make.sh
\
&&
./make.sh
install
\
&&
rm
-rf
${
FASTDFS_PATH
}
/libfastcommon
#compile the fastdfs
WORKDIR
${FASTDFS_PATH}/fastdfs
RUN
git clone
--branch
V5.11
--depth
1 https://github.com/happyfish100/fastdfs.git
${
FASTDFS_PATH
}
/fastdfs
\
&&
./make.sh
\
&&
./make.sh
install
\
&&
rm
-rf
${
FASTDFS_PATH
}
/fastdfs
EXPOSE
22122 23000 8080 8888
VOLUME
["$FASTDFS_BASE_PATH", "/etc/fdfs"]
COPY
conf/*.* /etc/fdfs/
COPY
start.sh /usr/bin/
#make the start.sh executable
RUN
chmod
777 /usr/bin/start.sh
ENTRYPOINT
["/usr/bin/start.sh"]
CMD
["tracker"]
baseimg/tools/fastdfs/conf/client.conf
0 → 100644
View file @
11371cdb
# connect timeout in seconds
# default value is 30s
connect_timeout
=
30
# network timeout in seconds
# default value is 30s
network_timeout
=
60
# the base path to store log files
base_path
=/
var
/
fdfs
# tracker_server can ocur more than once, and tracker_server format is
# "host:port", host can be hostname or ip address
tracker_server
=
192
.
168
.
0
.
197
:
22122
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level
=
info
# if use connection pool
# default value is false
# since V4.05
use_connection_pool
=
false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time
=
3600
# if load FastDFS parameters from tracker server
# since V4.05
# default value is false
load_fdfs_parameters_from_tracker
=
false
# if use storage ID instead of IP address
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# default value is false
# since V4.05
use_storage_id
=
false
# specify storage ids filename, can use relative or absolute path
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# since V4.05
storage_ids_filename
=
storage_ids
.
conf
#HTTP settings
http
.
tracker_server_port
=
80
#use "#include" directive to include HTTP other settiongs
##include http.conf
baseimg/tools/fastdfs/conf/http.conf
0 → 100644
View file @
11371cdb
# HTTP default content type
http
.
default_content_type
=
application
/
octet
-
stream
# MIME types mapping filename
# MIME types file format: MIME_type extensions
# such as: image/jpeg jpeg jpg jpe
# you can use apache's MIME file: mime.types
http
.
mime_types_filename
=
mime
.
types
# if use token to anti-steal
# default value is false (0)
http
.
anti_steal
.
check_token
=
false
# token TTL (time to live), seconds
# default value is 600
http
.
anti_steal
.
token_ttl
=
900
# secret key to generate anti-steal token
# this parameter must be set when http.anti_steal.check_token set to true
# the length of the secret key should not exceed 128 bytes
http
.
anti_steal
.
secret_key
=
FastDFS1234567890
# return the content of the file when check token fail
# default value is empty (no file sepecified)
http
.
anti_steal
.
token_check_fail
=/
home
/
yuqing
/
fastdfs
/
conf
/
anti
-
steal
.
jpg
baseimg/tools/fastdfs/conf/mime.types
0 → 100644
View file @
11371cdb
This diff is collapsed.
Click to expand it.
baseimg/tools/fastdfs/conf/storage.conf
0 → 100644
View file @
11371cdb
# is this config file disabled
# false for enabled
# true for disabled
disabled
=
false
# the name of the group this storage server belongs to
#
# comment or remove this item for fetching from tracker server,
# in this case, use_storage_id must set to true in tracker.conf,
# and storage_ids.conf must be configed correctly.
group_name
=
group1
# bind an address of this host
# empty for bind all addresses of this host
bind_addr
=
# if bind an address of this host when connect to other servers
# (this storage server as a client)
# true for binding the address configed by above parameter: "bind_addr"
# false for binding any address of this host
client_bind
=
true
# the storage server port
port
=
23000
# connect timeout in seconds
# default value is 30s
connect_timeout
=
30
# network timeout in seconds
# default value is 30s
network_timeout
=
60
# heart beat interval in seconds
heart_beat_interval
=
30
# disk usage report interval in seconds
stat_report_interval
=
60
# the base path to store data and log files
base_path
=/
var
/
fdfs
# max concurrent connections the server supported
# default value is 256
# more max_connections means more memory will be used
max_connections
=
256
# the buff size to recv / send data
# this parameter must more than 8KB
# default value is 64KB
# since V2.00
buff_size
=
256
KB
# accept thread count
# default value is 1
# since V4.07
accept_threads
=
1
# work thread count, should <= max_connections
# work thread deal network io
# default value is 4
# since V2.00
work_threads
=
4
# if disk read / write separated
## false for mixed read and write
## true for separated read and write
# default value is true
# since V2.00
disk_rw_separated
=
true
# disk reader thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_reader_threads
=
1
# disk writer thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_writer_threads
=
1
# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
sync_wait_msec
=
50
# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
sync_interval
=
0
# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time
=
00
:
00
# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_end_time
=
23
:
59
# write to the mark file after sync N files
# default value is 500
write_mark_file_freq
=
500
# path(disk or mount point) count, default value is 1
store_path_count
=
1
# store_path#, based 0, if store_path0 not exists, it's value is base_path
# the paths must be exist
store_path0
=/
var
/
fdfs
#store_path1=/var/fdfs2
# subdir_count * subdir_count directories will be auto created under each
# store_path (disk), value can be 1 to 256, default value is 256
subdir_count_per_path
=
256
# tracker_server can ocur more than once, and tracker_server format is
# "host:port", host can be hostname or ip address
tracker_server
=
192
.
168
.
209
.
121
:
22122
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level
=
info
#unix group name to run this program,
#not set (empty) means run by the group of current user
run_by_group
=
#unix username to run this program,
#not set (empty) means run by current user
run_by_user
=
# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts
=*
# the mode of the files distributed to the data path
# 0: round robin(default)
# 1: random, distributted by hash code
file_distribute_path_mode
=
0
# valid when file_distribute_to_path is set to 0 (round robin),
# when the written file count reaches this number, then rotate to next path
# default value is 100
file_distribute_rotate_count
=
100
# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
fsync_after_written_bytes
=
0
# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
sync_log_buff_interval
=
10
# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
sync_binlog_buff_interval
=
10
# sync storage stat info to disk every interval seconds
# default value is 300 seconds
sync_stat_file_interval
=
300
# thread stack size, should >= 512KB
# default value is 512KB
thread_stack_size
=
512
KB
# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
upload_priority
=
10
# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix
=
# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
check_file_duplicate
=
0
# file signature method for check file duplicate
## hash: four 32 bits hash code
## md5: MD5 signature
# default value is hash
# since V4.01
file_signature_method
=
hash
# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
key_namespace
=
FastDFS
# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
keep_alive
=
0
# you can use "#include filename" (not include double quotes) directive to
# load FastDHT server list, when the filename is a relative path such as
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf
# if log to access log
# default value is false
# since V4.00
use_access_log
=
false
# if rotate the access log every day
# default value is false
# since V4.00
rotate_access_log
=
false
# rotate access log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.00
access_log_rotate_time
=
00
:
00
# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log
=
false
# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time
=
00
:
00
# rotate access log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_access_log_size
=
0
# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size
=
0
# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days
=
0
# if skip the invalid record when sync file
# default value is false
# since V4.02
file_sync_skip_invalid_record
=
false
# if use connection pool
# default value is false
# since V4.05
use_connection_pool
=
false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time
=
3600
# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
http
.
domain_name
=
# the port of the web server on this storage server
http
.
server_port
=
8888
baseimg/tools/fastdfs/conf/storage_ids.conf
0 → 100644
View file @
11371cdb
# <id> <group_name> <ip_or_hostname>
# 100001 group1 192.168.0.196
# 100002 group1 192.168.0.116
baseimg/tools/fastdfs/conf/tracker.conf
0 → 100644
View file @
11371cdb
# is this config file disabled
# false for enabled
# true for disabled
disabled
=
false
# bind an address of this host
# empty for bind all addresses of this host
bind_addr
=
# the tracker server port
port
=
22122
# connect timeout in seconds
# default value is 30s
connect_timeout
=
30
# network timeout in seconds
# default value is 30s
network_timeout
=
60
# the base path to store data and log files
base_path
=/
var
/
fdfs
# max concurrent connections this server supported
max_connections
=
256
# accept thread count
# default value is 1
# since V4.07
accept_threads
=
1
# work thread count, should <= max_connections
# default value is 4
# since V2.00
work_threads
=
4
# min buff size
# default value 8KB
min_buff_size
=
8
KB
# max buff size
# default value 128KB
max_buff_size
=
128
KB
# the method of selecting group to upload files
# 0: round robin
# 1: specify group
# 2: load balance, select the max free space group to upload file
store_lookup
=
2
# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
store_group
=
group2
# which storage server to upload file
# 0: round robin (default)
# 1: the first server order by ip address
# 2: the first server order by priority (the minimal)
store_server
=
0
# which path(means disk or mount point) of the storage server to upload file
# 0: round robin
# 2: load balance, select the max free space path to upload file
store_path
=
0
# which storage server to download file
# 0: round robin (default)
# 1: the source storage server which the current file uploaded to
download_server
=
0
# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in
# a group <= reserved_storage_space,
# no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
### XX.XX% as ratio such as reserved_storage_space = 10%
reserved_storage_space
=
10
%
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level
=
info
#unix group name to run this program,
#not set (empty) means run by the group of current user
run_by_group
=
#unix username to run this program,
#not set (empty) means run by current user
run_by_user
=
# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts
=*
# sync log buff to disk every interval seconds
# default value is 10 seconds
sync_log_buff_interval
=
10
# check storage server alive interval seconds
check_active_interval
=
120
# thread stack size, should >= 64KB
# default value is 64KB
thread_stack_size
=
64
KB
# auto adjust when the ip address of the storage server changed
# default value is true
storage_ip_changed_auto_adjust
=
true
# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
storage_sync_file_max_delay
=
86400
# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
storage_sync_file_max_time
=
300
# if use a trunk file to store several small files
# default value is false
# since V3.00
use_trunk_file
=
false
# the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
slot_min_size
=
256
# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <= this value
# default value is 16MB
# since V3.00
slot_max_size
=
16
MB
# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
trunk_file_size
=
64
MB
# if create trunk file advancely
# default value is false
# since V3.06
trunk_create_file_advance
=
false
# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
# since V3.06
trunk_create_file_time_base
=
02
:
00
# the interval of create trunk file, unit: second
# default value is 38400 (one day)
# since V3.06
trunk_create_file_interval
=
86400
# the threshold to create trunk file
# when the free trunk file size less than the threshold, will create
# the trunk files
# default value is 0
# since V3.06
trunk_create_file_space_threshold
=
20
G
# if check trunk space occupying when loading trunk free spaces
# the occupied spaces will be ignored
# default value is false
# since V3.09
# NOTICE: set this parameter to true will slow the loading of trunk spaces
# when startup. you should set this parameter to true when neccessary.
trunk_init_check_occupying
=
false
# if ignore storage_trunk.dat, reload from trunk binlog
# default value is false
# since V3.10
# set to true once for version upgrade when your version less than V3.10
trunk_init_reload_from_binlog
=
false
# the min interval for compressing the trunk binlog file
# unit: second
# default value is 0, 0 means never compress
# FastDFS compress the trunk binlog when trunk init and trunk destroy
# recommand to set this parameter to 86400 (one day)
# since V5.01
trunk_compress_binlog_min_interval
=
0
# if use storage ID instead of IP address
# default value is false
# since V4.00
use_storage_id
=
false
# specify storage ids filename, can use relative or absolute path
# since V4.00
storage_ids_filename
=
storage_ids
.
conf
# id type of the storage server in the filename, values are:
## ip: the ip address of the storage server
## id: the server id of the storage server
# this paramter is valid only when use_storage_id set to true
# default value is ip
# since V4.03
id_type_in_filename
=
ip
# if store slave file use symbol link
# default value is false
# since V4.01
store_slave_file_use_link
=
false
# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log
=
false
# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time
=
00
:
00
# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size
=
0
# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days
=
0
# if use connection pool
# default value is false
# since V4.05
use_connection_pool
=
false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time
=
3600
# HTTP port on this tracker server
http
.
server_port
=
8080
# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http
.
check_alive_interval
=
30
# check storage HTTP server alive type, values are:
# tcp : connect to the storge server with HTTP port only,
# do not request and get response
# http: storage check alive url must return http status 200
# default value is tcp
http
.
check_alive_type
=
tcp
# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http
.
check_alive_uri
=/
status
.
html
baseimg/tools/fastdfs/start.sh
0 → 100644
View file @
11371cdb
#!/bin/bash
#set -e
if
[
"
$1
"
=
"monitor"
]
;
then
if
[
-n
"
$TRACKER_SERVER
"
]
;
then
sed
-i
"s|tracker_server=.*
$|
tracker_server=
${
TRACKER_SERVER
}
|g"
/etc/fdfs/client.conf
fi
fdfs_monitor /etc/fdfs/client.conf
exit
0
elif
[
"
$1
"
=
"storage"
]
;
then
FASTDFS_MODE
=
"storage"
else
FASTDFS_MODE
=
"tracker"
fi
if
[
-n
"
$PORT
"
]
;
then
sed
-i
"s|^port=.*
$|
port=
${
PORT
}
|g"
/etc/fdfs/
"
$FASTDFS_MODE
"
.conf
fi
if
[
-n
"
$TRACKER_SERVER
"
]
;
then
sed
-i
"s|tracker_server=.*
$|
tracker_server=
${
TRACKER_SERVER
}
|g"
/etc/fdfs/storage.conf
sed
-i
"s|tracker_server=.*
$|
tracker_server=
${
TRACKER_SERVER
}
|g"
/etc/fdfs/client.conf
fi
if
[
-n
"
$GROUP_NAME
"
]
;
then
sed
-i
"s|group_name=.*
$|
group_name=
${
GROUP_NAME
}
|g"
/etc/fdfs/storage.conf
fi
FASTDFS_LOG_FILE
=
"
${
FASTDFS_BASE_PATH
}
/logs/
${
FASTDFS_MODE
}
d.log"
PID_NUMBER
=
"
${
FASTDFS_BASE_PATH
}
/data/fdfs_
${
FASTDFS_MODE
}
d.pid"
echo
"try to start the
$FASTDFS_MODE
node..."
if
[
-f
"
$FASTDFS_LOG_FILE
"
]
;
then
rm
"
$FASTDFS_LOG_FILE
"
fi
# start the fastdfs node.
fdfs_
${
FASTDFS_MODE
}
d /etc/fdfs/
${
FASTDFS_MODE
}
.conf start
# wait for pid file(important!),the max start time is 5 seconds,if the pid number does not appear in 5 seconds,start failed.
TIMES
=
5
while
[
!
-f
"
$PID_NUMBER
"
-a
$TIMES
-gt
0
]
do
sleep
1s
TIMES
=
`
expr
$TIMES
- 1
`
done
# if the storage node start successfully, print the started time.
# if [ $TIMES -gt 0 ]; then
# echo "the ${FASTDFS_MODE} node started successfully at $(date +%Y-%m-%d_%H:%M)"
# # give the detail log address
# echo "please have a look at the log detail at $FASTDFS_LOG_FILE"
# # leave balnk lines to differ from next log.
# echo
# echo
# # make the container have foreground process(primary commond!)
# tail -F --pid=`cat $PID_NUMBER` /dev/null
# # else print the error.
# else
# echo "the ${FASTDFS_MODE} node started failed at $(date +%Y-%m-%d_%H:%M)"
# echo "please have a look at the log detail at $FASTDFS_LOG_FILE"
# echo
# echo
# fi
tail
-f
"
$FASTDFS_LOG_FILE
"
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment