docker-compse安装fastdfs

必须用 docker host网络模式进行通信

目录层级


├── docker-compose.yml
├── fastdfs.tar.gz
├── nginx
│   └── nginx.conf
├── storage
│   ├── conf
│   │   └── storage.conf
│   └── data
├── store_path
│   └── data
└── tracker├── conf│   ├── client.conf│   └── tracker.conf└── data

注意

配置 TRACKER_SERVER地址为虚机或公网服务器ip这里必须用host网络模式如果是虚机需要手动开启防火墙sudo firewall-cmd --zone=public --add-port=22122/tcp --permanentsudo firewall-cmd --zone=public --add-port=23000/tcp --permanentsudo firewall-cmd --zone=public --add-port=9800/tcp --permanentsudo firewall-cmd --reload

docker-compose.yml


version: "3.5"
services:fastdfs-tracker:hostname: fastdfs-trackercontainer_name: fastdfs-trackerimage: season/fastdfs:1.2network_mode: "host"command: trackervolumes:- ./tracker/data:/fastdfs/tracker/data- ./tracker/conf:/etc/fdfsfastdfs-storage:hostname: fastdfs-storagecontainer_name: fastdfs-storageimage: season/fastdfs:1.2network_mode: "host"volumes:- ./storage/data:/fastdfs/storage/data- ./store_path:/fastdfs/store_path- ./storage/conf/storage.conf:/fdfs_conf/storage.confenvironment:- TRACKER_SERVER=192.168.106.132:22122command: storagedepends_on:- fastdfs-trackerfastdfs-nginx:hostname: fastdfs-nginxcontainer_name: fastdfs-nginximage: season/fastdfs:1.2network_mode: "host"volumes:- ./nginx/nginx.conf:/etc/nginx/conf/nginx.conf- ./store_path:/fastdfs/store_pathenvironment:- TRACKER_SERVER=192.168.106.132:22122command: nginx

tracker.conf


# is this config file disabled
# false for enabled
# true for disabled
disabled=false# bind an address of this host
# empty for bind all addresses of this host
bind_addr=# the tracker server port
port=22122# connect timeout in seconds
# default value is 30s
connect_timeout=30# network timeout in seconds
# default value is 30s
network_timeout=60# the base path to store data and log files
base_path=/fastdfs/tracker# max concurrent connections this server supported
max_connections=256# accept thread count
# default value is 1
# since V4.07
accept_threads=1# work thread count, should <= max_connections
# default value is 4
# since V2.00
work_threads=4# the method of selecting group to upload files
# 0: round robin
# 1: specify group
# 2: load balance, select the max free space group to upload file
store_lookup=2# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
store_group=group1# which storage server to upload file
# 0: round robin (default)
# 1: the first server order by ip address
# 2: the first server order by priority (the minimal)
store_server=0# which path(means disk or mount point) of the storage server to upload file
# 0: round robin
# 2: load balance, select the max free space path to upload file
store_path=0# which storage server to download file
# 0: round robin (default)
# 1: the source storage server which the current file uploaded to
download_server=0# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in
# a group <= reserved_storage_space,
# no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
### XX.XX% as ratio such as reserved_storage_space = 10%
reserved_storage_space = 10%#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info#unix group name to run this program,
#not set (empty) means run by the group of current user
run_by_group=#unix username to run this program,
#not set (empty) means run by current user
run_by_user=# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
allow_hosts=*# sync log buff to disk every interval seconds
# default value is 10 seconds
sync_log_buff_interval = 10# check storage server alive interval seconds
check_active_interval = 120# thread stack size, should >= 64KB
# default value is 64KB
thread_stack_size = 64KB# auto adjust when the ip address of the storage server changed
# default value is true
storage_ip_changed_auto_adjust = true# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
storage_sync_file_max_delay = 86400# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
storage_sync_file_max_time = 300# if use a trunk file to store several small files
# default value is false
# since V3.00
use_trunk_file = false# the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
slot_min_size = 256# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <=  this value
# default value is 16MB
# since V3.00
slot_max_size = 16MB# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
trunk_file_size = 64MB# if create trunk file advancely
# default value is false
# since V3.06
trunk_create_file_advance = false# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
# since V3.06
trunk_create_file_time_base = 02:00# the interval of create trunk file, unit: second
# default value is 38400 (one day)
# since V3.06
trunk_create_file_interval = 86400# the threshold to create trunk file
# when the free trunk file size less than the threshold, will create
# the trunk files
# default value is 0
# since V3.06
trunk_create_file_space_threshold = 20G# if check trunk space occupying when loading trunk free spaces
# the occupied spaces will be ignored
# default value is false
# since V3.09
# NOTICE: set this parameter to true will slow the loading of trunk spaces
# when startup. you should set this parameter to true when neccessary.
trunk_init_check_occupying = false# if ignore storage_trunk.dat, reload from trunk binlog
# default value is false
# since V3.10
# set to true once for version upgrade when your version less than V3.10
trunk_init_reload_from_binlog = false# if use storage ID instead of IP address
# default value is false
# since V4.00
use_storage_id = false# specify storage ids filename, can use relative or absolute path
# since V4.00
storage_ids_filename = storage_ids.conf# id type of the storage server in the filename, values are:
## ip: the ip address of the storage server
## id: the server id of the storage server
# this paramter is valid only when use_storage_id set to true
# default value is ip
# since V4.03
id_type_in_filename = ip# if store slave file use symbol link
# default value is false
# since V4.01
store_slave_file_use_link = false# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time=00:00# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0# if use connection pool
# default value is false
# since V4.05
use_connection_pool = false# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600# HTTP port on this tracker server
http.server_port=8080# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http.check_alive_interval=30# check storage HTTP server alive type, values are:
#   tcp : connect to the storge server with HTTP port only,
#        do not request and get response
#   http: storage check alive url must return http status 200
# default value is tcp
http.check_alive_type=tcp# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http.check_alive_uri=/status.html

client.conf 这里修改ip为虚机或服务器


# connect timeout in seconds
# default value is 30s
connect_timeout=30
# network timeout in seconds
# default value is 30s
network_timeout=60# the base path to store log files
base_path=/fastdfs/client# tracker_server can ocur more than once, and tracker_server format is
#  "host:port", host can be hostname or ip address
# 需要修改此处 ip
tracker_server=192.168.106.132:22122#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info# if use connection pool
# default value is false
# since V4.05
use_connection_pool = false# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600# if load FastDFS parameters from tracker server
# since V4.05
# default value is false
load_fdfs_parameters_from_tracker=false# if use storage ID instead of IP address
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# default value is false
# since V4.05
use_storage_id = false# specify storage ids filename, can use relative or absolute path
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# since V4.05
storage_ids_filename = storage_ids.conf#HTTP settings
http.tracker_server_port=80#use "#include" directive to include HTTP other settiongs
##include http.conf

storage.conf 这里修改ip为虚机或服务器


# is this config file disabled
# false for enabled
# true for disabled
disabled=false# the name of the group this storage server belongs to
group_name=group1# bind an address of this host
# empty for bind all addresses of this host
bind_addr=# if bind an address of this host when connect to other servers 
# (this storage server as a client)
# true for binding the address configed by above parameter: "bind_addr"
# false for binding any address of this host
client_bind=true# the storage server port
port=23000# connect timeout in seconds
# default value is 30s
connect_timeout=30# network timeout in seconds
# default value is 30s
network_timeout=60# heart beat interval in seconds
heart_beat_interval=30# disk usage report interval in seconds
stat_report_interval=60# the base path to store data and log files
base_path=/fastdfs/storage# max concurrent connections the server supported
# default value is 256
# more max_connections means more memory will be used
max_connections=256# the buff size to recv / send data
# this parameter must more than 8KB
# default value is 64KB
# since V2.00
buff_size = 256KB# accept thread count
# default value is 1
# since V4.07
accept_threads=1# work thread count, should <= max_connections
# work thread deal network io
# default value is 4
# since V2.00
work_threads=4# if disk read / write separated
##  false for mixed read and write
##  true for separated read and write
# default value is true
# since V2.00
disk_rw_separated = true# disk reader thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_reader_threads = 1# disk writer thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_writer_threads = 1# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
sync_wait_msec=50# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
sync_interval=0# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time=00:00# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_end_time=23:59# write to the mark file after sync N files
# default value is 500
write_mark_file_freq=500# path(disk or mount point) count, default value is 1
store_path_count=1# store_path#, based 0, if store_path0 not exists, it's value is base_path
# the paths must be exist
store_path0=/fastdfs/store_path
#store_path1=/home/yuqing/fastdfs2# subdir_count  * subdir_count directories will be auto created under each 
# store_path (disk), value can be 1 to 256, default value is 256
subdir_count_per_path=256# tracker_server can ocur more than once, and tracker_server format is
#  "host:port", host can be hostname or ip address
tracker_server=192.168.106.132:22122#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level=info#unix group name to run this program, 
#not set (empty) means run by the group of current user
run_by_group=#unix username to run this program,
#not set (empty) means run by current user
run_by_user=# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
allow_hosts=*# the mode of the files distributed to the data path
# 0: round robin(default)
# 1: random, distributted by hash code
file_distribute_path_mode=0# valid when file_distribute_to_path is set to 0 (round robin), 
# when the written file count reaches this number, then rotate to next path
# default value is 100
file_distribute_rotate_count=100# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
fsync_after_written_bytes=0# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
sync_log_buff_interval=10# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
sync_binlog_buff_interval=10# sync storage stat info to disk every interval seconds
# default value is 300 seconds
sync_stat_file_interval=300# thread stack size, should >= 512KB
# default value is 512KB
thread_stack_size=512KB# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
upload_priority=10# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix=# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
check_file_duplicate=0# file signature method for check file duplicate
## hash: four 32 bits hash code
## md5: MD5 signature
# default value is hash
# since V4.01
file_signature_method=hash# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
key_namespace=FastDFS# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
keep_alive=0# you can use "#include filename" (not include double quotes) directive to 
# load FastDHT server list, when the filename is a relative path such as 
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf# if log to access log
# default value is false
# since V4.00
use_access_log = false# if rotate the access log every day
# default value is false
# since V4.00
rotate_access_log = false# rotate access log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.00
access_log_rotate_time=00:00# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time=00:00# rotate access log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_access_log_size = 0# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0# if skip the invalid record when sync file
# default value is false
# since V4.02
file_sync_skip_invalid_record=false# if use connection pool
# default value is false
# since V4.05
use_connection_pool = false# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
http.domain_name=# the port of the web server on this storage server
http.server_port=8888

java连接测试

<!--FastDFS客户端程序包--><dependency><groupId>net.oschina.zcx7878</groupId><artifactId>fastdfs-client-java</artifactId><version>1.27.0.0</version></dependency>

import org.csource.common.NameValuePair;
import org.csource.fastdfs.*;
import org.springframework.core.io.ClassPathResource;
import top.imuster.file.provider.file.FastDFSFile;import java.io.ByteArrayInputStream;
import java.io.InputStream;/*** 实现FastDFS文件管理*      文件上传*      文件删除*      文件下载*      文件信息获取*      Storage信息获取*      Tracker信息获取*/
public class FastDFSUtil {/*** @Description: 加载Tracker链接信息* @Author: lpf* @Date: 2019/12/23 12:29**/static {try {//查找classpath下的文件路径String fileNameConf = new ClassPathResource("fdfs_client.conf").getPath();//加载Tracker链接信息ClientGlobal.init(fileNameConf);} catch (Exception e) {e.printStackTrace();}}/*** @Description: 文件上传实现* @param fastDFSFile :上传的文件信息封装**/public static String[] upload(FastDFSFile fastDFSFile) throws Exception {//附加参数NameValuePair[] meta_list = new NameValuePair[1];meta_list[0] = new NameValuePair("author",fastDFSFile.getAuthor());//获取TrackerServerTrackerServer trackerServer = getTrackerServer();//获取StorageClientStorageClient storageClient = getStorageClient(trackerServer);/*** 通过StorageClient访问Storage,实现文件上传,并且获取文件上传后的存储信息* 函数的参数*      1.上传文件的字节数组*      2.文件的拓展名*      3.附加参数  比如拍摄地点:北京* uploads[]*      upload[0]:文件上传所存储的Storage的组的名字  group1*      upload[1]:文件存储到Storage上的文件的名字   M00/02/44/1.jpg**/String[] uploads = storageClient.upload_file(fastDFSFile.getContent(), fastDFSFile.getExt(), meta_list);return uploads;}/**** @Description: 获取文件信息* @param groupName :文件的组名   group1* @param remoteFileName :文件的存储路径名字  M00/02/44/1.jpg* @reture:**/public static FileInfo getFile(String groupName, String remoteFileName) throws Exception{//获取TrackerServerTrackerServer trackerServer = getTrackerServer();//获取StorageClientStorageClient storageClient = getStorageClient(trackerServer);//获取文件信息return storageClient.get_file_info(groupName, remoteFileName);}/*** @Description: 文件下载* @param groupName :文件的组名   group1* @param remoteFileName :文件的存储路径名字  M00/02/44/1.jpg* @reture: void**/public static InputStream downloadFile(String groupName, String remoteFileName) throws Exception{//获取TrackerServerTrackerServer trackerServer = getTrackerServer();//获取StorageClientStorageClient storageClient = getStorageClient(trackerServer);//文件下载byte[] buffer = storageClient.download_file(groupName, remoteFileName);return new ByteArrayInputStream(buffer);}/*** @Description: 删除文件* @param groupName :文件的组名   group1* @param remoteFileName :文件的存储路径名字  M00/02/44/1.jpg* @reture: void**/public static void deleteFile(String groupName, String remoteFileName) throws Exception{//获取TrackerServerTrackerServer trackerServer = getTrackerServer();//获取StorageClientStorageClient storageClient = getStorageClient(trackerServer);//删除文件storageClient.delete_file(groupName, remoteFileName);}/*** @Description: 获取storage信息* @reture: 返回storage信息**/public static StorageServer getStorage() throws Exception{//创建一个TrackerClient对象,通过TrackerClient对象访问TrackerServerTrackerClient trackerClient = new TrackerClient();//通过TrackerClient获取TrackerServer的链接对象TrackerServer trackerServer = trackerClient.getConnection();//获取storage信息return trackerClient.getStoreStorage(trackerServer);}/*** @Description: 获取storage信息的端口和IP信息* @reture: 返回storage的IP和端口信息**/public static ServerInfo[] getServerInfo (String groupName, String remoteFileName) throws Exception{//创建一个TrackerClient对象,通过TrackerClient对象访问TrackerServerTrackerClient trackerClient = new TrackerClient();//通过TrackerClient获取TrackerServer的链接对象TrackerServer trackerServer = trackerClient.getConnection();//获取Storage的IP和端口信息return trackerClient.getFetchStorages(trackerServer, groupName, groupName);}/*** @Description: 获取Tracker的信息* @param* @reture: void**/public static String getTrackerInfo() throws Exception{//获取TrackerServerTrackerServer trackerServer = getTrackerServer();//Tracker的IP,Http端口String ip = trackerServer.getInetSocketAddress().getHostString();int httpPort = ClientGlobal.getG_tracker_http_port();String url = "http://" + ip + ":" + httpPort;return url;}/*** @Description: 获取TrackerServer(防止冗余)* @param* @reture: org.csource.fastdfs.TrackerServer**/public static TrackerServer getTrackerServer() throws Exception{//创建一个TrackerClient对象,通过TrackerClient对象访问TrackerServerTrackerClient trackerClient = new TrackerClient();//通过TrackerClient获取TrackerServer的链接对象TrackerServer trackerServer = trackerClient.getConnection();return trackerServer;}/*** @Description: 获取StorageClient* @param trackerServer* @reture: org.csource.fastdfs.TrackerServer**/public static StorageClient getStorageClient(TrackerServer trackerServer) throws Exception {StorageClient storageClient = new StorageClient(trackerServer,null);return storageClient;}//测试public static void main(String[] args) throws Exception{/*FileInfo fileInfo = getFile("group1", "M00/00/00/rBgYGV4AkH-AafqxAAGGegCgxtc834.jpg");System.out.println(fileInfo.getSourceIpAddr());System.out.println(fileInfo.getFileSize());*//* //文件下载InputStream is = downloadFile("group1", "M00/00/00/rBgYGV4Ag3aACWPGAAIsTF3RcOs943.jpg");//将文件写入到本地磁盘FileOutputStream os = new FileOutputStream("D:\\tool\\images\\test_download\\1.jpg");//定义一个缓冲区byte[] buffer = new byte[1024];while (is.read(buffer) != -1) {os.write(buffer);}os.flush();  //刷新缓冲区os.close();is.close();*///测试文件删除//deleteFile("group1", "M00/00/00/rBgYGV4AijSAD9iqAAIVMmsTSos650.jpg");//获取storage信息/*StorageServer storageServer = getStorage();System.out.println(storageServer.getStorePathIndex());System.out.println(storageServer.getInetSocketAddress().getHostString());   //IP信息*///获取Storage组的IP和端口信息/*ServerInfo[] groups = getServerInfo("group1", "M00/00/00/rBgYGV4AkH-AafqxAAGGegCgxtc834.jpg");System.out.println(groups);for (ServerInfo group : groups) {System.out.println(group.getIpAddr());System.out.println(group.getPort());}*///获取tracker的信息//System.out.println(getTrackerInfo());}
}
public static void main(String[] args) {try{//封装文件信息File file = new File("D:\\1.png");FileInputStream inputStream = new FileInputStream(file);byte[] bytes = IoUtil.readBytes(inputStream);FastDFSFile fastDFSFile = new FastDFSFile("1.png",   //文件名 1.jpgbytes,              //文件的字节数组org.springframework.util.StringUtils.getFilenameExtension("1.png")    //获取文件拓展名);//调用FastDFSUtil工具类将文件上传到FastDFS中String[] uploads = FastDFSUtil.upload(fastDFSFile);//拼接访问地址 url = http://192.168.106.132:9800/group1/M00/00/00/CgAQCmYp93GARnrvAAAWcPkOFKA731.png//String url = "http://192.168.106.132:9800/" + uploads[0] + "/" + uploads[1];Message<Object> bySuccess = Message.createBySuccess(uploads[0] + "/" + uploads[1]);System.out.println(uploads[0] + "/" + uploads[1]);}catch (Exception e){e.printStackTrace();}}

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处:http://www.mzph.cn/diannao/4577.shtml

如若内容造成侵权/违法违规/事实不符,请联系多彩编程网进行投诉反馈email:809451989@qq.com,一经查实,立即删除!

相关文章

MySQL使用Sequence创建唯一主键

目录 第一章、快速了解Sequence1.1&#xff09;是什么&#xff1f;为什么使用1.2&#xff09;Sequence和自增主键的区别 第二章、在MySQL中使用Sequence2.1&#xff09;创建mysql_sequence表2.1.1&#xff09;创建表2.1.2&#xff09;插入数据 2.2&#xff09;创建函数2.2.1&am…

vue使用debugger调试代码

1.在vue.config.js中 在你的vue文件vue.config.js中找到如下configureWebpack位置&#xff0c;加入devtool:"source map"&#xff0c;重新启动项目debugger即可生效。

分享:9.3版本无缝导入AVEVA PDMS高版本工程12.0,12.1,E3D

9.3版本可以无缝导入AVEVA PDMS的工程。 UKP3d导入AVEVA PDMS工程的方法 http://47.94.91.234/forum.php?modviewthread&tid163583&fromuid6 (出处: 优易软件-工厂设计软件专家) &#xff08;从AVEVA PDMS导出时元件和等级的功能我们正做收尾工作&#xff0c;到时可以…

如何进行制造设备数据汇集,发挥数据的价值?

数字化转型正深刻推动制造企业实现远程监控、提高生产效率、降低生产成本、优化产品质量及明晰精细化方向。并且工业互联网的发展离不开工业数据的应用&#xff0c;而制造设备数据汇集正是应用的基础。但制造设备数据汇集存在以下难点及痛点&#xff1a; 1、安全把控难 关键的…

Mysql基础(一)DDL、DML、DQL、DCL 扫盲之DDL语句

一 扫盲 SQL语句根据其功能主要分为四类&#xff1a;DDL、DML、DQL、DCL说明&#xff1a; 本篇章只会粗略的分类,不会展开细节讲解 1、DDL&#xff08;Data Definition Language&#xff09;说明&#xff1a; 是一组用于定义和管理数据库结构的语句2、DML&#xff08;Data Ma…

RLDP协议原理与应用

RLDP概述 l RLDP全称是Rapid Link Detection Protocol&#xff08;快速链路检测协议&#xff09;&#xff0c;是锐捷网络自主开发的&#xff0c;用于快速检测以太网链路故障的链路协议。 l 一般的以太网链路检测机制都只是利用物理连接的状态&#xff0c;通过物理层的自动协…

张大哥笔记:我付钱了,我就是大爷?

很抱歉用这个当做标题&#xff0c;来给大家分享一些电商的故事&#xff01;大家好&#xff0c;我是张大哥&#xff0c;今天聊聊在电商路上遇到过的奇葩买家&#xff1f; 比如最近我在做PDD的时候&#xff0c;就会遇到很多莫名其妙的sha子&#xff0c;咱是知识份子&#xff0c;肯…

远程桌面连接不上个别服务器的问题分析与解决方案

在日常的IT运维工作中&#xff0c;远程桌面连接&#xff08;RDP&#xff0c;Remote Desktop Protocol&#xff09;是我们经常使用的工具之一&#xff0c;用于管理和维护远程服务器。然而&#xff0c;有时我们可能会遇到无法连接到个别服务器的情况。针对这一问题&#xff0c;我…

【算法模板】数学基础算法模版

文章目录 数学基础试除法判定质数试除法分解质因数朴素筛法求素数线性筛法求素数试除法求所有约数约数个数和约数之和欧几里得算法求欧拉函数筛法求欧拉函数快速幂扩展欧几里得算法 高斯消元解线性方程组递推法求组合数通过预处理逆元的方式求组合数 Lucas定理分解质因数法求组…

微信小程序4~6章总结

目录 第四章 页面组件总结 4.1 组件的定义及属性 4.2 容器视图组件 4.2.1 view 4.2.2 scroll-view 4.2.3 swiper 4.3 基础内容组件 4.3.1 icon ​编辑 4.3.2 text 4.3.3 progress ​编辑 4.4 表单组件 4.4.1 button 4.4.2 radio 4.4.3 checkbox 4.4.4 switch …

C语言--贪吃蛇小游戏

目录 一、Win32API介绍 1.1Win32API 1.2控制台程序 1.3控制台屏幕上的坐标COORD 1.4GetStdHandle 1.5GetConsoleCursorInfo 1.6 CONSOLE_CURSOR_INFO 1.7 SetConsoleCursorInfo 1.8SetConsoleCursorPosition 1.9GetAsyncKeyState 二、贪吃蛇游戏设计与分析 2.1地图 …

手机群控操作软件分析

随着移动互联网的快速发展,智能手机已经成为人们生活中不可或缺的一部分。无论是工作、学习还是娱乐,手机都扮演着重要的角色。然而,随着手机数量的不断增加,对于大量手机的集中管理和操作变得愈发困难。为了解决这一问题,手机群控操作软件应运而生。本文将对手机群控操作…

网站内容下载软件有哪些 网站内容下载软件推荐 网站内容下载软件安全吗 idm是啥软件 idm网络下载免费

一招搞定网页内容下载&#xff0c;并且各大网站通用&#xff01;绕过资源审查&#xff0c;所有网站内容随意下载。解锁速度限制&#xff0c;下载即高速无视网站限速。跳过会员充值&#xff0c;所有VIP资源免费下载。有关网站内容下载软件有哪些&#xff0c;网站内容下载软件推荐…

【leetcode】快慢指针相关题目总结

141. 环形链表 判断链表是否有环&#xff1a;如果链表中存在环&#xff0c;则在链表上不断前进的指针会一直在环里绕圈子&#xff0c;且不能知道链表是否有环。使用快慢指针&#xff0c;当链表中存在环时&#xff0c;两个指针最终会在环中相遇。 /*** Definition for singly-…

代谢组数据分析四:功能分析

Functional Analysis 代谢物通路包含了基因、催化酶或代谢物等上下游关系的先验知识,通过将关心的代谢物比对到通路上,再根据如超级几何分析等数学方法计算受影响的代谢物是否能够影响通路。 代谢物富集分析的目的是为了解析某些差异代谢物是否落在某些pathway上(可简单理…

Ubuntu-22.04电源选项设置

提示&#xff1a;文章写完后&#xff0c;目录可以自动生成&#xff0c;如何生成可参考右边的帮助文档 文章目录 前言说明一、图形设置二、命令行设置1.查看支持的模式2.设置新模式 总结 前言 我们使用Windows都知道电源选项有省电、平衡和高性能模式。其实Ubuntu-22.04也有这个…

【Linux】文件系统

送给大家一句话&#xff1a; 你的任务&#xff0c;就是珍惜你自己的人生&#xff0c;而且还要比之前任何时候更加珍惜。 – 东野圭吾 文件系统 1 前言2 物理磁盘3 磁盘的存储结构4 抽象理解磁盘储存5 引入文件系统 &#xff08;如何管理磁盘文件&#xff09;5.1 了解文件系统5.…

基于 SpringCloud 的在线交易平台乐优商城的设计与实现(六)

目录 第六章 系统测试 6.1 功能性测试 6.1.1 商家后台功能测试 6.1.2 前台功能测试 6.2 非功能性测试 6.3 本章小结 结束语 参考文献 前面内容请移步 基于 SpringCloud 的在线交易平台乐优商城的设计与实现&#xff08;五&#xff09; 相关免费源码资源 乐优商城…

SpringBoot JPA使用

SpringBoot JPA使用 简介 Spring Data JPA 是 Spring 框架提供的一个模块&#xff0c;用于简化与关系型数据库的交互和数据访问。它基于JPA&#xff08;Java Persistence API&#xff09;标准&#xff0c;并提供了一组易于使用的API和工具&#xff0c;帮助开发人员更轻松地进…

深入理解分布式事务① ---->分布式事务基础(四大特性、五大类型、本地事务、MySQL并发事务问题、MySQL事务隔离级别命令设置)详解

目录 深入理解分布式事务① ---->分布式事务基础&#xff08;四大特性、五大类型、本地事务、MySQL并发事务问题、MySQL事务隔离级别命令设置&#xff09;详解事务的基本概念1、什么是事务&#xff1f;2、事务的四大特性2-1&#xff1a;原子性&#xff08;Atomic&#xff09…