add
This commit is contained in:
30
23-浙江交投/1-base-env-shell/0.0-dependencies.sh
Normal file
30
23-浙江交投/1-base-env-shell/0.0-dependencies.sh
Normal file
@@ -0,0 +1,30 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
# 需要在所有的节点执行
|
||||
|
||||
hostnamectl set-hostname master-node
|
||||
|
||||
sed -i "/search/ a nameserver 223.5.5.5" /etc/resolv.conf
|
||||
|
||||
echo "AllowTcpForwarding yes" >> /etc/ssh/sshd_config
|
||||
systemctl restart sshd
|
||||
|
||||
cat >> /etc/hosts << EOF
|
||||
10.100.2.121 master-node
|
||||
10.100.2.122 worker-1
|
||||
10.100.2.116 storage-1
|
||||
EOF
|
||||
|
||||
ping worker-1
|
||||
|
||||
yum clean all && yum makecache
|
||||
|
||||
|
||||
root
|
||||
|
||||
10.100.2.121
|
||||
|
||||
SuperCyy.123
|
||||
|
||||
|
||||
|
||||
95
23-浙江交投/1-base-env-shell/0.1-mountNodeVolume.sh
Normal file
95
23-浙江交投/1-base-env-shell/0.1-mountNodeVolume.sh
Normal file
@@ -0,0 +1,95 @@
|
||||
#! /bin/bash
|
||||
|
||||
# 关闭虚拟缓存
|
||||
swapoff -a
|
||||
cp -f /etc/fstab /etc/fstab_bak
|
||||
cat /etc/fstab_bak | grep -v swap >/etc/fstab
|
||||
|
||||
# echo "-----------------------------------------------------------------------"
|
||||
# RootVolumeSizeBefore=$(df -TH | grep -w "/dev/mapper/centos-root" | awk '{print $3}')
|
||||
# echo "扩容之前的root目录的容量为:${RootVolumeSizeBefore}"
|
||||
|
||||
# echo "y
|
||||
|
||||
|
||||
# " | lvremove /dev/mapper/centos-swap
|
||||
|
||||
# freepesize=$(vgdisplay centos | grep 'Free PE' | awk '{print $5}')
|
||||
|
||||
# lvextend -l+${freepesize} /dev/mapper/centos-root
|
||||
|
||||
|
||||
# ## #自动扩展XFS文件系统到最大的可用大小
|
||||
# xfs_growfs /dev/mapper/centos-root
|
||||
|
||||
# df -TH | grep -w "/dev/mapper/centos-root" | awk '{print $3}'
|
||||
|
||||
# echo "-----------------------------------------------------------------------"
|
||||
# RootVolumeSizeAfter=$(df -TH | grep -w "/dev/mapper/centos-root" | awk '{print $3}')
|
||||
# echo "扩容之后的root目录的容量为:${RootVolumeSizeAfter}"
|
||||
# RootVolumeSizeBeforeNum=$(echo $RootVolumeSizeBefore | cut -d "G" -f1)
|
||||
# RootVolumeSizeAfterNum=$(echo $RootVolumeSizeAfter | cut -d "G" -f1)
|
||||
|
||||
# echo "恭喜,您的root目录容量增加了+++++++$(( ${RootVolumeSizeAfterNum}-${RootVolumeSizeBeforeNum} ))GB+++++"
|
||||
|
||||
|
||||
yum install lvm2 -y
|
||||
echo ""
|
||||
echo ""
|
||||
echo ""
|
||||
echo "-----------------------------------------------------------------------"
|
||||
|
||||
export VG_NAME=datavg
|
||||
|
||||
echo "n
|
||||
p
|
||||
|
||||
|
||||
|
||||
t
|
||||
|
||||
8e
|
||||
w
|
||||
" | fdisk /dev/sda
|
||||
partprobe
|
||||
# 如果已经存在卷组,直接进行添加
|
||||
# vgextend /dev/mapper/centos /dev/vda3
|
||||
vgcreate ${VG_NAME} /dev/vdb1
|
||||
export selfpesize=$(vgdisplay ${VG_NAME} | grep 'Total PE' | awk '{print $3}')
|
||||
# 大小根据实际情况调整
|
||||
lvcreate -l ${selfpesize} -n lvdata ${VG_NAME}
|
||||
mkfs.xfs /dev/mapper/${VG_NAME}-lvdata
|
||||
mkdir -p /data
|
||||
mkdir -p /var/lib/docker
|
||||
#selffstab="/dev/mapper/${VG_NAME}-lvdata /var/lib/docker xfs defaults 0 0"
|
||||
export selffstab="/dev/mapper/${VG_NAME}-lvdata /var/lib/docker xfs defaults 0 0"
|
||||
echo "${selffstab}" >> /etc/fstab
|
||||
mount -a
|
||||
|
||||
echo ""
|
||||
echo ""
|
||||
echo ""
|
||||
df -TH
|
||||
echo "-----------------------------------------------------------------------"
|
||||
|
||||
# 扩容根目录,${VG_NAME}-root 通过df -Th获取需要扩容的文件系统
|
||||
# lvextend -l +100%FREE /dev/mapper/${VG_NAME}-root
|
||||
# xfs_growfs /dev/mapper/${VG_NAME}-root
|
||||
|
||||
# 自定义 安装lvm2'
|
||||
echo "n
|
||||
p
|
||||
|
||||
|
||||
|
||||
t
|
||||
|
||||
8e
|
||||
w
|
||||
" | fdisk /dev/vda
|
||||
partprobe
|
||||
vgextend klas_host-10-190-202-141 /dev/vda4
|
||||
lvextend -l +100%FREE /dev/mapper/klas_host--10--190--202--141-root
|
||||
partprobe
|
||||
xfs_growfs /dev/mapper/klas_host--10--190--202--141-root
|
||||
df -TH
|
||||
2929
23-浙江交投/1-base-env-shell/1_node_important.sh
Normal file
2929
23-浙江交投/1-base-env-shell/1_node_important.sh
Normal file
File diff suppressed because it is too large
Load Diff
147
23-浙江交投/1-base-env-shell/2-imageDownSync.sh
Normal file
147
23-浙江交投/1-base-env-shell/2-imageDownSync.sh
Normal file
@@ -0,0 +1,147 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
### 需要修改以下的内容 ###
|
||||
#### 需要修改以下的内容 ###
|
||||
#### 需要修改以下的内容 ###
|
||||
|
||||
cmlc_app_image_list="cmlc-app-images-4.1.6.txt" # 需要修改版本
|
||||
rancher_image_list="kubernetes-images-2.5.7-1.20.4.txt" # 一般不需要修改
|
||||
middleware_image_list="middleware-images.txt" # 一般不需要修改
|
||||
#DockerRegisterDomain="20.47.129.116:8033" # 需要根据实际修改
|
||||
DockerRegisterDomain="172.10.125.92:8033" # 需要根据实际修改
|
||||
HarborAdminPass=V2ryStr@ngPss # 需要跟第一脚本中的密码保持一致
|
||||
|
||||
#### 需要修改以上的内容 ###
|
||||
#### 需要修改以上的内容 ###
|
||||
#### 需要修改以上的内容 ###
|
||||
|
||||
downloadAllNeededImages() {
|
||||
while [[ $# > 0 ]]; do
|
||||
pulled=""
|
||||
while IFS= read -r i; do
|
||||
[ -z "${i}" ] && continue
|
||||
echo "开始下载:${i}"
|
||||
if docker pull "${i}" >/dev/null 2>&1; then
|
||||
echo "Image pull success: ${i}"
|
||||
pulled="${pulled} ${i}"
|
||||
else
|
||||
if docker inspect "${i}" >/dev/null 2>&1; then
|
||||
pulled="${pulled} ${i}"
|
||||
else
|
||||
echo "Image pull failed: ${i}"
|
||||
fi
|
||||
fi
|
||||
echo "-------------------------------------------------"
|
||||
done <"${1}"
|
||||
shift
|
||||
done
|
||||
}
|
||||
|
||||
downloadAllNeededImagesAndCompress() {
|
||||
while [[ $# > 0 ]]; do
|
||||
pulled=""
|
||||
while IFS= read -r i; do
|
||||
[ -z "${i}" ] && continue
|
||||
echo "开始下载:${i}"
|
||||
if docker pull "${i}" >/dev/null 2>&1; then
|
||||
echo "Image pull success: ${i}"
|
||||
pulled="${pulled} ${i}"
|
||||
else
|
||||
if docker inspect "${i}" >/dev/null 2>&1; then
|
||||
pulled="${pulled} ${i}"
|
||||
else
|
||||
echo "Image pull failed: ${i}"
|
||||
fi
|
||||
fi
|
||||
echo "-------------------------------------------------"
|
||||
done <"${1}"
|
||||
compressPacName="$(echo ${1} | cut -d"." -f1).tar.gz"
|
||||
|
||||
echo "Creating ${compressPacName} with $(echo ${pulled} | wc -w | tr -d '[:space:]') images"
|
||||
docker save $(echo ${pulled}) | gzip --stdout > ${compressPacName}
|
||||
|
||||
shift
|
||||
done
|
||||
|
||||
|
||||
echo "已经完成打包工作!"
|
||||
}
|
||||
|
||||
pushRKEImageToHarbor(){
|
||||
linux_images=()
|
||||
while IFS= read -r i; do
|
||||
[ -z "${i}" ] && continue
|
||||
linux_images+=("${i}");
|
||||
done < "${rancher_image_list}"
|
||||
|
||||
docker login -u admin -p ${HarborAdminPass} ${DockerRegisterDomain}
|
||||
|
||||
for i in "${linux_images[@]}"; do
|
||||
[ -z "${i}" ] && continue
|
||||
case $i in
|
||||
*/*)
|
||||
image_name="${DockerRegisterDomain}/${i}"
|
||||
;;
|
||||
*)
|
||||
image_name="${DockerRegisterDomain}/rancher/${i}"
|
||||
;;
|
||||
esac
|
||||
|
||||
echo "开始镜像至私有仓库推送:${image_name}"
|
||||
docker tag "${i}" "${image_name}"
|
||||
docker push "${image_name}"
|
||||
echo "-------------------------------------------------"
|
||||
done
|
||||
}
|
||||
|
||||
pushCMLCAPPImageToHarbor(){
|
||||
app_images=()
|
||||
while IFS= read -r i; do
|
||||
[ -z "${i}" ] && continue
|
||||
app_images+=("${i}");
|
||||
done < "${cmlc_app_image_list}"
|
||||
|
||||
docker login -u admin -p ${HarborAdminPass} ${DockerRegisterDomain}
|
||||
for app in "${app_images[@]}"; do
|
||||
[ -z "${app}" ] && continue
|
||||
image_name="${DockerRegisterDomain}/$(echo ${app} | cut -d"/" -f2-8)"
|
||||
echo "开始镜像至私有仓库推送:${image_name}"
|
||||
docker tag "${app}" "${image_name}"
|
||||
docker push "${image_name}"
|
||||
echo "-------------------------------------------------"
|
||||
done
|
||||
}
|
||||
|
||||
pushMiddlewareImageToHarbor(){
|
||||
middleware_image=()
|
||||
while IFS= read -r i; do
|
||||
[ -z "${i}" ] && continue
|
||||
middleware_image+=("${i}");
|
||||
done < "${middleware_image_list}"
|
||||
|
||||
docker login -u admin -p ${HarborAdminPass} ${DockerRegisterDomain}
|
||||
for app in "${middleware_image[@]}"; do
|
||||
[ -z "${app}" ] && continue
|
||||
case ${app} in
|
||||
*/*/*)
|
||||
image_name="${DockerRegisterDomain}/cmii/$(echo "${app}" | cut -d"/" -f3-8)"
|
||||
;;
|
||||
*/*)
|
||||
image_name="${DockerRegisterDomain}/cmii/$(echo "${app}" | cut -d"/" -f2-8)"
|
||||
;;
|
||||
esac
|
||||
|
||||
echo "开始镜像至私有仓库推送:${image_name}"
|
||||
docker tag "${app}" "${image_name}"
|
||||
docker push "${image_name}"
|
||||
echo "-------------------------------------------------"
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
#downloadAllNeededImagesAndCompress "${middleware_image_list}"
|
||||
downloadAllNeededImages "${rancher_image_list}"
|
||||
|
||||
#pushRKEImageToHarbor
|
||||
#pushCMLCAPPImageToHarbor
|
||||
#pushMiddlewareImageToHarbor
|
||||
237
23-浙江交投/1-base-env-shell/3-bootUPk8s.sh
Normal file
237
23-浙江交投/1-base-env-shell/3-bootUPk8s.sh
Normal file
@@ -0,0 +1,237 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
### 需要修改以下的内容 ###
|
||||
### 需要修改以下的内容 ###
|
||||
### 需要修改以下的内容 ###
|
||||
|
||||
# 理论上来说,能访问公网的服务器 用来部署Harbor服务器
|
||||
# 所有的主机均可以访问公网的话,填写 除了harbor服务器的 其他所有主机的地址
|
||||
PrivateServerIPs=(10.215.125.15 10.215.125.16 10.215.125.17) # 内网服务器的IP地址,不包括可以访问公网IP的服务器
|
||||
|
||||
### 需要修改以上的内容 ###
|
||||
### 需要修改以上的内容 ###
|
||||
### 需要修改以上的内容 ###
|
||||
|
||||
RED="31m" ## 姨妈红
|
||||
GREEN="32m" ## 水鸭青
|
||||
YELLOW="33m" ## 鸭屎黄
|
||||
PURPLE="35m" ## 基佬紫
|
||||
BLUE="36m" ## 天依蓝
|
||||
|
||||
colorEcho() {
|
||||
# shellcheck disable=SC2145
|
||||
echo -e "\033[${1}${@:2}\033[0m" 1>&2
|
||||
}
|
||||
|
||||
check_root() {
|
||||
if [[ $EUID != 0 ]]; then
|
||||
colorEcho ${RED} "当前非root账号(或没有root权限),无法继续操作,请更换root账号!"
|
||||
colorEcho ${YELLOW} "使用sudo -命令获取临时root权限(执行后可能会提示输入root密码)"
|
||||
exit 1
|
||||
fi
|
||||
}
|
||||
|
||||
startFunc(){
|
||||
colorEcho ${PURPLE} "---------------------------------------------------------------------------------"
|
||||
colorEcho ${BLUE} "开始执行 启动RKE集群的操作 ………"
|
||||
echo ""
|
||||
|
||||
colorEcho ${BLUE} "本脚本的运行有一些列的前提依赖,请确定以下的项目都已完成!!!!"
|
||||
colorEcho ${YELLOW} "----------------------------------------------------------"
|
||||
colorEcho ${RED} "1. 完成基础环境初始化,将rke系统镜像均上传至私有Harbor中!"
|
||||
colorEcho ${RED} "2. 配置并修改好 rke集群的模板文件,命名为 cluster.yml !!"
|
||||
colorEcho ${RED} "3. ……"
|
||||
colorEcho ${YELLOW} "----------------------------------------------------------"
|
||||
|
||||
while true; do
|
||||
colorEcho ${RED} "请确保您已经将上述的项目完成!!"
|
||||
read -r -p "请输入yes进行确认,脚本才可继续运行!!" input
|
||||
case $input in
|
||||
yes)
|
||||
colorEcho ${GREEN} "您已确认上述的项目均已完成!!"
|
||||
colorEcho ${GREEN} "----------------------------------------------------------"
|
||||
echo ""
|
||||
colorEcho ${BLUE} "开始执行 RKE集群的启动过程!!!"
|
||||
echo ""
|
||||
main
|
||||
break
|
||||
;;
|
||||
*)
|
||||
echo ""
|
||||
colorEcho ${RED} "输入有误!!! 请输入 >> yes << 进行确认"
|
||||
break
|
||||
colorEcho ${RED} "-----------------------------------------------------"
|
||||
echo ""
|
||||
;;
|
||||
esac
|
||||
done
|
||||
}
|
||||
|
||||
installRKE(){
|
||||
colorEcho ${PURPLE} "---------------------------------------------------------------------------------"
|
||||
colorEcho ${BLUE} "开始下载并安装 RKE 工具 ………"
|
||||
echo ""
|
||||
|
||||
colorEcho ${BLUE} "开始从rancher镜像下载rke工具……"
|
||||
wget http://rancher-mirror.cnrancher.com/rke/v1.2.6/rke_linux-amd64
|
||||
if [ -s rke_linux-amd64 ]; then
|
||||
colorEcho ${GREEN} "rke工具下载完成!"
|
||||
chmod +x rke_linux-amd64
|
||||
mv ./rke_linux-amd64 /usr/local/bin/rke
|
||||
colorEcho ${GREEN} "----------------------------------------------------------"
|
||||
rke --version
|
||||
colorEcho ${GREEN} "----------------------------------------------------------"
|
||||
rke config --list-version --all
|
||||
echo ""
|
||||
colorEcho ${BLUE} "开始从rancher镜像下载 kubectl 工具……"
|
||||
wget http://rancher-mirror.cnrancher.com/kubectl/v1.20.4/linux-amd64-v1.20.4-kubectl
|
||||
chmod +x linux-amd64-v1.20.4-kubectl
|
||||
mv linux-amd64-v1.20.4-kubectl /usr/local/bin/kubectl
|
||||
colorEcho ${GREEN} "----------------------------------------------------------"
|
||||
kubectl version
|
||||
colorEcho ${GREEN} "----------------------------------------------------------"
|
||||
else
|
||||
colorEcho ${RED} "rke工具下载失败!!!脚本无法继续运行,请手动下载rke工具!"
|
||||
colorEcho ${RED} "rke工具下载失败!!!脚本无法继续运行,请手动下载rke工具!"
|
||||
colorEcho ${RED} "rke工具下载失败!!!脚本无法继续运行,请手动下载rke工具!"
|
||||
return 1
|
||||
fi
|
||||
}
|
||||
|
||||
createRKEInstallerUser(){
|
||||
colorEcho ${PURPLE} "---------------------------------------------------------------------------------"
|
||||
colorEcho ${BLUE} "开始创建 rke-installer 用户………"
|
||||
echo ""
|
||||
useradd rke-installer
|
||||
echo "SuperCyy.123
|
||||
SuperCyy.123" | passwd rke-installer
|
||||
#将登陆用户develop加入到docker用户组中
|
||||
gpasswd -a rke-installer docker
|
||||
#更新用户组
|
||||
newgrp docker
|
||||
echo ""
|
||||
if [ -d /home/rke-installer ]; then
|
||||
colorEcho ${GREEN} "rke-installer 用户创建成功!! "
|
||||
echo ""
|
||||
else
|
||||
mkdir -p /home/rke-installer
|
||||
chown rke-installer:rke-installer -R /home/rke-installer
|
||||
usermod -d /home/rke-installer rke-installer
|
||||
colorEcho ${YELLOW} "检测到 rke-installer 用户已经存在"
|
||||
fi
|
||||
|
||||
if [[ -s cluster.yaml || -s cluster.yml ]]; then
|
||||
colorEcho ${BLUE} "开始将 cluster.yaml文件复制到 rke-installer目录下…………"
|
||||
mv cluster.y* /home/rke-installer/cluster.yml
|
||||
if [ -s /home/rke-installer/cluster.yml ]; then
|
||||
colorEcho ${BLUE} "cluster.yml文件已经放置完成!"
|
||||
chown rke-installer:rke-installer /home/rke-installer/cluster.yml
|
||||
else
|
||||
colorEcho ${RED} "当前目录下未检测到 rke集群的模板文件!!"
|
||||
colorEcho ${RED} "程序无法继续,将退出!!"
|
||||
return 1
|
||||
fi
|
||||
else
|
||||
colorEcho ${RED} "当前目录下未检测到 rke集群的模板文件!!"
|
||||
colorEcho ${RED} "程序无法继续,将退出!!"
|
||||
echo ""
|
||||
colorEcho ${YELLOW} "--------------------------------------------------"
|
||||
colorEcho ${RED} "请创建RKE集群的模板文件,并命名为 cluster.yml "
|
||||
colorEcho ${RED} "请创建RKE集群的模板文件,并命名为 cluster.yml "
|
||||
colorEcho ${RED} "请创建RKE集群的模板文件,并命名为 cluster.yml "
|
||||
colorEcho ${YELLOW} "--------------------------------------------------"
|
||||
return 1
|
||||
fi
|
||||
|
||||
colorEcho ${BLUE} "开始切换当前用户至 rke-installer "
|
||||
su rke-installer
|
||||
echo ""
|
||||
colorEcho ${BLUE} "请检查rke-installer用户能否执行 docker ps 命令!!"
|
||||
docker ps
|
||||
colorEcho ${BLUE} "----------------------------------------------------------"
|
||||
}
|
||||
|
||||
generateRKEUserKey(){
|
||||
colorEcho ${PURPLE} "---------------------------------------------------------------------------------"
|
||||
colorEcho ${BLUE} "开始创建 rke-installer用户的 ssh key ……"
|
||||
echo ""
|
||||
su rke-installer
|
||||
ssh-keygen -t rsa -P "" -f ~/.ssh/id_rsa
|
||||
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
|
||||
chmod 600 ~/.ssh/authorized_keys
|
||||
colorEcho ${GREEN} "--------------------------------------------------------------"
|
||||
colorEcho ${GREEN} "-----------本机配置完成!-------------"
|
||||
echo ""
|
||||
|
||||
for ip in "${PrivateServerIPs[@]}"; do
|
||||
colorEcho ${BLUE} "请手动将如下的命令,以 root 权限在主机 ${ip} 上运行"
|
||||
colorEcho ${BLUE} "请手动将如下的命令,以 root 权限在主机 ${ip} 上运行"
|
||||
colorEcho ${BLUE} "请手动将如下的命令,以 root 权限在主机 ${ip} 上运行"
|
||||
colorEcho ${BLUE} "-----------------------------------------------"
|
||||
echo ""
|
||||
echo ""
|
||||
colorEcho ${RED} " 请以 root 角色 运行!!! "
|
||||
colorEcho ${RED} " 请以 root 角色 运行!!! "
|
||||
colorEcho ${RED} " 请以 root 角色 运行!!! "
|
||||
echo ""
|
||||
colorEcho ${YELLOW} "useradd rke-installer && echo \"rke-installer
|
||||
rke-installer
|
||||
\" | passwd rke-installer && gpasswd -a rke-installer docker && newgrp docker && su rke-installer && docker ps "
|
||||
echo ""
|
||||
colorEcho ${YELLOW} "clear && ssh-keygen -t rsa -P \"\" -f ~/.ssh/id_rsa && echo \"$(cat ~/.ssh/id_rsa.pub)\" >> ~/.ssh/authorized_keys && echo \"\" && cat ~/.ssh/authorized_keys"
|
||||
echo ""
|
||||
echo ""
|
||||
while true; do
|
||||
colorEcho ${RED} "请确保您已经将上述的命令在主机${ip}上执行了!!"
|
||||
read -r -p "请输入yes进行确认,脚本才可继续运行!!" input
|
||||
case $input in
|
||||
yes)
|
||||
colorEcho ${GREEN} "您已确认在主机${ip}上添加了私有的ssh key!"
|
||||
echo ""
|
||||
break
|
||||
;;
|
||||
*)
|
||||
echo ""
|
||||
colorEcho ${RED} "输入有误!!! 请输入 >> yes << 进行确认"
|
||||
colorEcho ${RED} "请在主机${ip}上执行上述命令!!!"
|
||||
colorEcho ${RED} "否则本脚本的功能会失效!!"
|
||||
colorEcho ${RED} "-----------------------------------------------------"
|
||||
echo ""
|
||||
;;
|
||||
esac
|
||||
done
|
||||
done
|
||||
}
|
||||
|
||||
startRKECLuster(){
|
||||
colorEcho ${PURPLE} "---------------------------------------------------------------------------------"
|
||||
colorEcho ${BLUE} "开始 启动 rke集群 !!!"
|
||||
colorEcho ${BLUE} "开始 启动 rke集群 !!!"
|
||||
colorEcho ${BLUE} "开始 启动 rke集群 !!!"
|
||||
echo ""
|
||||
if [[ $(pwd) == "/home/rke-installer" ]]; then
|
||||
colorEcho ${BLUE} "检测到当前目录为 /home/rke-installer"
|
||||
echo ""
|
||||
colorEcho ${BLUE} "开始执行 RKE 集群的启动过程 !!!"
|
||||
colorEcho ${BLUE} "-------------------------------------------------------------"
|
||||
for i in {3..1..-1}; do
|
||||
colorEcho ${BLUE} "倒计时开始 ->> $i 秒 <<-,准备启动RKE,上文的日志输出将会消失!!"
|
||||
sleep 2
|
||||
done
|
||||
clear
|
||||
rke up
|
||||
else
|
||||
colorEcho ${BLUE} "当前目录不为 /home/rke-installer,开始跳转目录!"
|
||||
cd /home/rke-installer
|
||||
startRKECLuster
|
||||
fi
|
||||
}
|
||||
|
||||
main(){
|
||||
check_root
|
||||
installRKE || return $?
|
||||
generateRKEUserKey || return $?
|
||||
startRKECLuster || return $?
|
||||
}
|
||||
|
||||
startFunc
|
||||
62
23-浙江交投/1-base-env-shell/cmlc-app-images-4.1.6.txt
Normal file
62
23-浙江交投/1-base-env-shell/cmlc-app-images-4.1.6.txt
Normal file
@@ -0,0 +1,62 @@
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-data:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-user:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-open-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervision:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervisionh5:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-supervision:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-airspace:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-alarm:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-autowaypoint:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-brain:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cloud-live:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-clusters:3.3.3
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cms:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-data-post-process:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-depotautoreturn:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-developer:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-device:4.1.6-0913
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-emergency:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gis-server:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-datasource:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-engine:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-manage:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-integration:4.1.6-0913
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-kpi-monitor:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-logger:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-material-warehouse:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mission:4.1.6-0913
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mqtthandler:4.1.6-0913
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-notice:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-oauth:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-ai-brain:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-armypeople:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-base:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-cms-portal:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-detection:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-emergency-rescue:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-logistics:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-media:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-multiterminal:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-mws:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-oms:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-open:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-security:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-securityh5:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-share:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-splice:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-visualization:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-process:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:4.1.6-0913
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-user:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-waypoint:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-live-operator:v4.0.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-rtsp-operator:v4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/zlm-mediaserver:v1.0.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-srs-oss-adaptor:v4.0.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-srs-operator:v4.0.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:4.1.0
|
||||
61
23-浙江交投/1-base-env-shell/cmlc-app-images-5.0.0.txt
Normal file
61
23-浙江交投/1-base-env-shell/cmlc-app-images-5.0.0.txt
Normal file
@@ -0,0 +1,61 @@
|
||||
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-data:4.1.6-test
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-user:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-app-release:4.1.6-20006
|
||||
minio/minio:RELEASE.2022-10-24T18-35-07Z
|
||||
harbor.cdcyy.com.cn/cmii/cmii-open-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervision:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervisionh5:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-supervision:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-airspace:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-alarm:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-autowaypoint:4.1.6-cm-0828
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-brain:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cloud-live:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cms:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-data-post-process:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-developer:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-device:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-emergency:4.1.6-beta
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gateway:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gis-server:4.1.6-24241
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-datasource:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-engine:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-manage:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-kpi-monitor:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-logger:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-material-warehouse:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mission:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mqtthandler:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-notice:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-oauth:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform:4.1.6-24241
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-ai-brain:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-armypeople:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-base:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-cms-portal:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-detection:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-emergency-rescue:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-logistics:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-media:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-multiterminal:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-mws:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-oms:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-open:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-security:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-securityh5:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-share:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-splice:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-visualization:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-process:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-user:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-waypoint:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-srs-operator:v4.0.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-live-operator:v4.0.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-rtsp-operator:v4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/zlm-mediaserver:v1.0.6
|
||||
harbor.cdcyy.com.cn/cmii/nfs-subdir-external-provisioner:v4.0.2
|
||||
69
23-浙江交投/1-base-env-shell/cmlc-app-images-5.1.0.txt
Normal file
69
23-浙江交投/1-base-env-shell/cmlc-app-images-5.1.0.txt
Normal file
@@ -0,0 +1,69 @@
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-data:5.1.0-staging
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-data:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-gateway:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-admin-user:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-app-release:4.2.0-validation
|
||||
harbor.cdcyy.com.cn/cmii/cmii-open-gateway:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervision:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervisionh5:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-suav-supervision:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-airspace:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-alarm:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-autowaypoint:4.1.6-cm-0828
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-brain:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cloud-live:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-clusters:3.3.3
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-cms:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-data-post-process:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-depotautoreturn:4.1.6-beta
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-developer:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-device:5.1.0-staging
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-device:5.1.0-24469-1012
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-emergency:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gateway:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-gis-server:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-datasource:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-engine:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-grid-manage:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:5.1.0-staging
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-integration:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-kpi-monitor:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-logger:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-material-warehouse:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mission:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mission:5.1.0-pr107
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-mqtthandler:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-notice:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-oauth:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform:5.1.0-24469-1012
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-ai-brain:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-armypeople:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-base:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-cms-portal:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-detection:5.0.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-emergency-rescue:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform:4.2.0-hljtt
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-logistics:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-media:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-multiterminal:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-mws:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-oms:5.1.0-24092
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-open:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-security:4.1.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-securityh5:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-share:5.1.0-24092
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-splice:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-platform-visualization:4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-process:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:5.1.0-24469-1012
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-tower:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-user:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-uav-waypoint:5.1.0
|
||||
harbor.cdcyy.com.cn/cmii/cmii-live-operator:v4.0.6
|
||||
harbor.cdcyy.com.cn/cmii/cmii-rtsp-operator:v4.1.0
|
||||
harbor.cdcyy.com.cn/cmii/zlm-mediaserver:v1.0.6
|
||||
harbor.cdcyy.com.cn/cmii/nfs-subdir-external-provisioner:v4.0.2
|
||||
44
23-浙江交投/1-base-env-shell/harbor.yml
Normal file
44
23-浙江交投/1-base-env-shell/harbor.yml
Normal file
@@ -0,0 +1,44 @@
|
||||
hostname: 20.47.129.116
|
||||
|
||||
http:
|
||||
port: 8033
|
||||
|
||||
harbor_admin_password: V2ryStr@ngPss
|
||||
|
||||
database:
|
||||
password: V2ryStr@ngPss
|
||||
max_idle_conns: 50
|
||||
max_open_conns: 1000
|
||||
|
||||
data_volume: /var/lib/docker/harbor-data
|
||||
|
||||
clair:
|
||||
updaters_interval: 12
|
||||
|
||||
jobservice:
|
||||
max_job_workers: 10
|
||||
|
||||
notification:
|
||||
webhook_job_max_retry: 10
|
||||
|
||||
chart:
|
||||
absolute_url: disabled
|
||||
|
||||
log:
|
||||
level: error
|
||||
local:
|
||||
rotate_count: 50
|
||||
rotate_size: 200M
|
||||
location: /var/log/harbor
|
||||
|
||||
_version: 2.0.0
|
||||
|
||||
proxy:
|
||||
http_proxy:
|
||||
https_proxy:
|
||||
no_proxy:
|
||||
components:
|
||||
- core
|
||||
- jobservice
|
||||
- clair
|
||||
- trivy
|
||||
64
23-浙江交投/1-base-env-shell/kubernetes-images-2.5.7-1.18.6.txt
Normal file
64
23-浙江交投/1-base-env-shell/kubernetes-images-2.5.7-1.18.6.txt
Normal file
@@ -0,0 +1,64 @@
|
||||
busybox
|
||||
rancher/backup-restore-operator:v1.0.3
|
||||
rancher/calico-cni:v3.13.4
|
||||
rancher/calico-ctl:v3.13.4
|
||||
rancher/calico-kube-controllers:v3.13.4
|
||||
rancher/calico-node:v3.13.4
|
||||
rancher/calico-pod2daemon-flexvol:v3.13.4
|
||||
rancher/cis-operator:v1.0.3
|
||||
rancher/cluster-proportional-autoscaler:1.7.1
|
||||
rancher/configmap-reload:v0.3.0-rancher4
|
||||
rancher/coredns-coredns:1.6.9
|
||||
rancher/coreos-etcd:v3.4.3-rancher1
|
||||
rancher/coreos-flannel:v0.12.0
|
||||
rancher/coreos-flannel:v0.13.0-rancher1
|
||||
rancher/coreos-kube-state-metrics:v1.9.7
|
||||
rancher/coreos-prometheus-config-reloader:v0.39.0
|
||||
rancher/coreos-prometheus-operator:v0.39.0
|
||||
rancher/externalip-webhook:v0.1.6
|
||||
rancher/flannel-cni:v0.3.0-rancher6
|
||||
rancher/fleet-agent:v0.3.4
|
||||
rancher/fleet:v0.3.4
|
||||
rancher/fluentd:v0.1.24
|
||||
rancher/grafana-grafana:7.1.5
|
||||
rancher/hyperkube:v1.18.16-rancher1
|
||||
rancher/jimmidyson-configmap-reload:v0.3.0
|
||||
rancher/k8s-dns-dnsmasq-nanny:1.15.2
|
||||
rancher/k8s-dns-kube-dns:1.15.2
|
||||
rancher/k8s-dns-node-cache:1.15.7
|
||||
rancher/k8s-dns-sidecar:1.15.2
|
||||
rancher/klipper-lb:v0.1.2
|
||||
rancher/kube-api-auth:v0.1.4
|
||||
rancher/kubectl:v1.18.6
|
||||
rancher/kubernetes-external-dns:v0.7.3
|
||||
rancher/library-busybox:1.31.1
|
||||
rancher/library-busybox:1.32.1
|
||||
rancher/library-nginx:1.19.2-alpine
|
||||
rancher/library-traefik:1.7.19
|
||||
rancher/local-path-provisioner:v0.0.11
|
||||
rancher/local-path-provisioner:v0.0.14
|
||||
rancher/local-path-provisioner:v0.0.19
|
||||
rancher/log-aggregator:v0.1.7
|
||||
rancher/istio-kubectl:1.5.10
|
||||
rancher/metrics-server:v0.3.6
|
||||
rancher/nginx-ingress-controller-defaultbackend:1.5-rancher1
|
||||
rancher/nginx-ingress-controller:nginx-0.35.0-rancher2
|
||||
rancher/opa-gatekeeper:v3.1.0-beta.7
|
||||
rancher/openzipkin-zipkin:2.14.2
|
||||
rancher/pause:3.1
|
||||
rancher/plugins-docker:18.09
|
||||
rancher/prom-alertmanager:v0.21.0
|
||||
rancher/prom-node-exporter:v1.0.1
|
||||
rancher/prom-prometheus:v2.12.0
|
||||
rancher/prom-prometheus:v2.18.2
|
||||
rancher/prometheus-auth:v0.2.1
|
||||
rancher/rancher-agent:v2.5.7
|
||||
rancher/rancher-webhook:v0.1.0-beta9
|
||||
rancher/rancher:v2.5.7
|
||||
rancher/rke-tools:v0.1.72
|
||||
rancher/security-scan:v0.1.14
|
||||
rancher/security-scan:v0.2.2
|
||||
rancher/shell:v0.1.6
|
||||
rancher/sonobuoy-sonobuoy:v0.16.3
|
||||
rancher/system-upgrade-controller:v0.6.2
|
||||
|
||||
63
23-浙江交投/1-base-env-shell/kubernetes-images-2.5.7-1.20.4.txt
Normal file
63
23-浙江交投/1-base-env-shell/kubernetes-images-2.5.7-1.20.4.txt
Normal file
@@ -0,0 +1,63 @@
|
||||
busybox
|
||||
rancher/backup-restore-operator:v1.0.3
|
||||
rancher/calico-cni:v3.17.2
|
||||
rancher/calico-ctl:v3.17.2
|
||||
rancher/calico-kube-controllers:v3.17.2
|
||||
rancher/calico-node:v3.17.2
|
||||
rancher/calico-pod2daemon-flexvol:v3.17.2
|
||||
rancher/cis-operator:v1.0.3
|
||||
rancher/cluster-proportional-autoscaler:1.7.1
|
||||
rancher/coredns-coredns:1.8.0
|
||||
rancher/coreos-etcd:v3.4.14-rancher1
|
||||
rancher/coreos-kube-state-metrics:v1.9.7
|
||||
rancher/coreos-prometheus-config-reloader:v0.39.0
|
||||
rancher/coreos-prometheus-operator:v0.39.0
|
||||
rancher/externalip-webhook:v0.1.6
|
||||
rancher/flannel-cni:v0.3.0-rancher6
|
||||
rancher/coreos-flannel:v0.13.0-rancher1
|
||||
rancher/fleet-agent:v0.3.4
|
||||
rancher/fleet:v0.3.4
|
||||
rancher/fluentd:v0.1.24
|
||||
rancher/grafana-grafana:7.1.5
|
||||
rancher/hyperkube:v1.20.4-rancher1
|
||||
rancher/jimmidyson-configmap-reload:v0.3.0
|
||||
rancher/k8s-dns-dnsmasq-nanny:1.15.2
|
||||
rancher/k8s-dns-kube-dns:1.15.2
|
||||
rancher/k8s-dns-node-cache:1.15.13
|
||||
rancher/k8s-dns-sidecar:1.15.2
|
||||
rancher/klipper-lb:v0.1.2
|
||||
rancher/kube-api-auth:v0.1.4
|
||||
rancher/kubectl:v1.20.4
|
||||
rancher/kubernetes-external-dns:v0.7.3
|
||||
rancher/cluster-proportional-autoscaler:1.8.1
|
||||
rancher/library-busybox:1.31.1
|
||||
rancher/library-busybox:1.32.1
|
||||
rancher/library-nginx:1.19.2-alpine
|
||||
rancher/library-traefik:1.7.19
|
||||
rancher/local-path-provisioner:v0.0.11
|
||||
rancher/local-path-provisioner:v0.0.14
|
||||
rancher/local-path-provisioner:v0.0.19
|
||||
rancher/log-aggregator:v0.1.7
|
||||
rancher/istio-kubectl:1.5.10
|
||||
rancher/metrics-server:v0.4.1
|
||||
rancher/configmap-reload:v0.3.0-rancher4
|
||||
rancher/nginx-ingress-controller-defaultbackend:1.5-rancher1
|
||||
rancher/nginx-ingress-controller:nginx-0.43.0-rancher1
|
||||
rancher/opa-gatekeeper:v3.1.0-beta.7
|
||||
rancher/openzipkin-zipkin:2.14.2
|
||||
rancher/pause:3.2
|
||||
rancher/plugins-docker:18.09
|
||||
rancher/prom-alertmanager:v0.21.0
|
||||
rancher/prom-node-exporter:v1.0.1
|
||||
rancher/prom-prometheus:v2.18.2
|
||||
rancher/prometheus-auth:v0.2.1
|
||||
rancher/rancher-agent:v2.5.7
|
||||
rancher/rancher-webhook:v0.1.0-beta9
|
||||
rancher/rancher:v2.5.7
|
||||
rancher/rke-tools:v0.1.72
|
||||
rancher/security-scan:v0.1.14
|
||||
rancher/security-scan:v0.2.2
|
||||
rancher/shell:v0.1.6
|
||||
rancher/sonobuoy-sonobuoy:v0.16.3
|
||||
rancher/system-upgrade-controller:v0.6.2
|
||||
|
||||
24
23-浙江交投/1-base-env-shell/middleware-images.txt
Normal file
24
23-浙江交投/1-base-env-shell/middleware-images.txt
Normal file
@@ -0,0 +1,24 @@
|
||||
docker.io/bitnami/redis:6.2.6-debian-10-r0
|
||||
docker.io/bitnami/redis:6.2.14-debian-11-r1
|
||||
docker.io/bitnami/mysql:8.0.35-debian-11-r1
|
||||
docker.io/bitnami/mysql:8.1.0-debian-11-r42
|
||||
docker.io/simonrupf/chronyd:0.4.3
|
||||
docker.io/bitnami/bitnami-shell:10-debian-10-r140
|
||||
docker.io/bitnami/bitnami-shell:11-debian-11-r136
|
||||
docker.io/bitnami/rabbitmq:3.9.12-debian-10-r3
|
||||
docker.io/bitnami/rabbitmq:3.11.26-debian-11-r2
|
||||
docker.io/ossrs/srs:v4.0.136
|
||||
docker.io/emqx/emqx:4.2.12
|
||||
docker.io/nacos/nacos-server:v2.1.2
|
||||
docker.io/nacos/nacos-server:v2.1.2-slim
|
||||
docker.io/mongo:5.0
|
||||
docker.io/rabbitmq:3.9-management
|
||||
docker.io/bitnami/minio:2022.5.4
|
||||
docker.io/bitnami/minio:2023.5.4
|
||||
docker.io/kubernetesui/dashboard:v2.0.1
|
||||
docker.io/kubernetesui/metrics-scraper:v1.0.4
|
||||
docker.io/ossrs/srs:v4.0-r3
|
||||
docker.io/nginx:1.21.3
|
||||
docker.io/redis:6.0.20-alpine
|
||||
docker.io/dyrnq/nfs-subdir-external-provisioner:v4.0.2
|
||||
|
||||
217
23-浙江交投/1-base-env-shell/rke-cluster.yml
Normal file
217
23-浙江交投/1-base-env-shell/rke-cluster.yml
Normal file
@@ -0,0 +1,217 @@
|
||||
nodes:
|
||||
- address: 20.47.129.116
|
||||
user: root
|
||||
role:
|
||||
- controlplane
|
||||
- etcd
|
||||
- worker
|
||||
internal_address: 20.47.129.116
|
||||
labels:
|
||||
ingress-deploy: true
|
||||
- address: 20.47.129.117
|
||||
user: root
|
||||
role:
|
||||
- worker
|
||||
internal_address: 20.47.129.117
|
||||
- address: 20.47.129.118
|
||||
user: root
|
||||
role:
|
||||
- worker
|
||||
internal_address: 20.47.129.118
|
||||
labels:
|
||||
mysql-deploy: true
|
||||
- address: 20.47.129.119
|
||||
user: root
|
||||
role:
|
||||
- worker
|
||||
internal_address: 20.47.129.119
|
||||
labels:
|
||||
mysql-deploy: true
|
||||
|
||||
|
||||
authentication:
|
||||
strategy: x509
|
||||
sans:
|
||||
- "20.47.129.116"
|
||||
|
||||
private_registries:
|
||||
- url: 20.47.129.116:8033 # 私有镜像库地址
|
||||
user: admin
|
||||
password: "V2ryStr@ngPss"
|
||||
is_default: true
|
||||
|
||||
##############################################################################
|
||||
|
||||
# 默认值为false,如果设置为true,当发现不支持的Docker版本时,RKE不会报错
|
||||
ignore_docker_version: true
|
||||
|
||||
# Set the name of the Kubernetes cluster
|
||||
cluster_name: rke-cluster
|
||||
|
||||
kubernetes_version: v1.20.4-rancher1-1
|
||||
|
||||
#ssh_key_path: /root/.ssh/id_ed25519
|
||||
ssh_key_path: /root/.ssh/id_rsa
|
||||
|
||||
# Enable running cri-dockerd
|
||||
# Up to Kubernetes 1.23, kubelet contained code called dockershim
|
||||
# to support Docker runtime. The replacement is called cri-dockerd
|
||||
# and should be enabled if you want to keep using Docker as your
|
||||
# container runtime
|
||||
# Only available to enable in Kubernetes 1.21 and higher
|
||||
enable_cri_dockerd: true
|
||||
|
||||
services:
|
||||
etcd:
|
||||
backup_config:
|
||||
enabled: false
|
||||
interval_hours: 72
|
||||
retention: 3
|
||||
safe_timestamp: false
|
||||
timeout: 300
|
||||
creation: 12h
|
||||
extra_args:
|
||||
election-timeout: 5000
|
||||
heartbeat-interval: 500
|
||||
gid: 0
|
||||
retention: 72h
|
||||
snapshot: false
|
||||
uid: 0
|
||||
|
||||
kube-api:
|
||||
# IP range for any services created on Kubernetes
|
||||
# This must match the service_cluster_ip_range in kube-controller
|
||||
service_cluster_ip_range: 172.24.0.0/16
|
||||
# Expose a different port range for NodePort services
|
||||
service_node_port_range: 30000-40000
|
||||
always_pull_images: true
|
||||
pod_security_policy: false
|
||||
# Add additional arguments to the kubernetes API server
|
||||
# This WILL OVERRIDE any existing defaults
|
||||
extra_args:
|
||||
# Enable audit log to stdout
|
||||
audit-log-path: "-"
|
||||
# Increase number of delete workers
|
||||
delete-collection-workers: 3
|
||||
# Set the level of log output to warning-level
|
||||
v: 1
|
||||
kube-controller:
|
||||
# CIDR pool used to assign IP addresses to pods in the cluster
|
||||
cluster_cidr: 172.28.0.0/16
|
||||
# IP range for any services created on Kubernetes
|
||||
# This must match the service_cluster_ip_range in kube-api
|
||||
service_cluster_ip_range: 172.24.0.0/16
|
||||
# Add additional arguments to the kubernetes API server
|
||||
# This WILL OVERRIDE any existing defaults
|
||||
extra_args:
|
||||
# Set the level of log output to debug-level
|
||||
v: 1
|
||||
# Enable RotateKubeletServerCertificate feature gate
|
||||
feature-gates: RotateKubeletServerCertificate=true
|
||||
# Enable TLS Certificates management
|
||||
# https://kubernetes.io/docs/tasks/tls/managing-tls-in-a-cluster/
|
||||
cluster-signing-cert-file: "/etc/kubernetes/ssl/kube-ca.pem"
|
||||
cluster-signing-key-file: "/etc/kubernetes/ssl/kube-ca-key.pem"
|
||||
kubelet:
|
||||
# Base domain for the cluster
|
||||
cluster_domain: cluster.local
|
||||
# IP address for the DNS service endpoint
|
||||
cluster_dns_server: 172.24.0.10
|
||||
# Fail if swap is on
|
||||
fail_swap_on: false
|
||||
# Set max pods to 250 instead of default 110
|
||||
extra_binds:
|
||||
- "/data/minio-pv:/hostStorage" # 不要修改 为minio的pv添加
|
||||
extra_args:
|
||||
max-pods: 122
|
||||
# Optionally define additional volume binds to a service
|
||||
scheduler:
|
||||
extra_args:
|
||||
# Set the level of log output to warning-level
|
||||
v: 0
|
||||
kubeproxy:
|
||||
extra_args:
|
||||
# Set the level of log output to warning-level
|
||||
v: 1
|
||||
|
||||
authorization:
|
||||
mode: rbac
|
||||
|
||||
addon_job_timeout: 30
|
||||
|
||||
# Specify network plugin-in (canal, calico, flannel, weave, or none)
|
||||
network:
|
||||
options:
|
||||
flannel_backend_type: vxlan
|
||||
flannel_iface: ens160
|
||||
flannel_autoscaler_priority_class_name: system-cluster-critical # Available as of RKE v1.2.6+
|
||||
flannel_priority_class_name: system-cluster-critical # Available as of RKE v1.2.6+
|
||||
plugin: flannel
|
||||
|
||||
# Specify DNS provider (coredns or kube-dns)
|
||||
dns:
|
||||
provider: coredns
|
||||
nodelocal: {}
|
||||
# Available as of v1.1.0
|
||||
update_strategy:
|
||||
strategy: RollingUpdate
|
||||
rollingUpdate:
|
||||
maxUnavailable: 20%
|
||||
maxSurge: 15%
|
||||
linear_autoscaler_params:
|
||||
cores_per_replica: 0.34
|
||||
nodes_per_replica: 4
|
||||
prevent_single_point_failure: true
|
||||
min: 2
|
||||
max: 3
|
||||
|
||||
# Specify monitoring provider (metrics-server)
|
||||
monitoring:
|
||||
provider: metrics-server
|
||||
# Available as of v1.1.0
|
||||
update_strategy:
|
||||
strategy: RollingUpdate
|
||||
rollingUpdate:
|
||||
maxUnavailable: 8
|
||||
|
||||
ingress:
|
||||
provider: nginx
|
||||
default_backend: true
|
||||
http_port: 0
|
||||
https_port: 0
|
||||
extra_envs:
|
||||
- name: TZ
|
||||
value: Asia/Shanghai
|
||||
node_selector:
|
||||
ingress-deploy: true
|
||||
options:
|
||||
use-forwarded-headers: "true"
|
||||
access-log-path: /var/log/nginx/access.log
|
||||
client-body-timeout: '6000'
|
||||
compute-full-forwarded-for: 'true'
|
||||
enable-underscores-in-headers: 'true'
|
||||
log-format-escape-json: 'true'
|
||||
log-format-upstream: >-
|
||||
{ "msec": "$msec", "connection": "$connection", "connection_requests":
|
||||
"$connection_requests", "pid": "$pid", "request_id": "$request_id",
|
||||
"request_length": "$request_length", "remote_addr": "$remote_addr",
|
||||
"remote_user": "$remote_user", "remote_port": "$remote_port",
|
||||
"http_x_forwarded_for": "$http_x_forwarded_for", "time_local":
|
||||
"$time_local", "time_iso8601": "$time_iso8601", "request": "$request",
|
||||
"request_uri": "$request_uri", "args": "$args", "status": "$status",
|
||||
"body_bytes_sent": "$body_bytes_sent", "bytes_sent": "$bytes_sent",
|
||||
"http_referer": "$http_referer", "http_user_agent": "$http_user_agent",
|
||||
"http_host": "$http_host", "server_name": "$server_name", "request_time":
|
||||
"$request_time", "upstream": "$upstream_addr", "upstream_connect_time":
|
||||
"$upstream_connect_time", "upstream_header_time": "$upstream_header_time",
|
||||
"upstream_response_time": "$upstream_response_time",
|
||||
"upstream_response_length": "$upstream_response_length",
|
||||
"upstream_cache_status": "$upstream_cache_status", "ssl_protocol":
|
||||
"$ssl_protocol", "ssl_cipher": "$ssl_cipher", "scheme": "$scheme",
|
||||
"request_method": "$request_method", "server_protocol": "$server_protocol",
|
||||
"pipe": "$pipe", "gzip_ratio": "$gzip_ratio", "http_cf_ray": "$http_cf_ray",
|
||||
"geoip_country_code": "$geoip_country_code" }
|
||||
proxy-body-size: 5120m
|
||||
proxy-read-timeout: '6000'
|
||||
proxy-send-timeout: '6000'
|
||||
|
||||
80
23-浙江交投/1-base-env-shell/常用脚本.sh
Normal file
80
23-浙江交投/1-base-env-shell/常用脚本.sh
Normal file
@@ -0,0 +1,80 @@
|
||||
# 下载所有离线文件
|
||||
|
||||
wget https://oss.demo.uavcmlc.com:18000/cmlc-installation/kebite-4.1.6.tar.gz
|
||||
|
||||
wget https://oss.demo.uavcmlc.com:18000/cmlc-installation/v4.1.6/middleware-images.tar.gz
|
||||
wget https://oss.demo.uavcmlc.com:18000/cmlc-installation/v4.1.6/rancher-1.20.4-image.tar.gz
|
||||
|
||||
|
||||
wget https://oss.demo.uavcmlc.com:18000/cmlc-installation/v4.1.6/rke
|
||||
|
||||
|
||||
# 批量复制文件
|
||||
ip_list=(10.20.1.133 10.20.1.134 10.20.1.132)
|
||||
for ip in "${ip_list[@]}"; do
|
||||
|
||||
echo "yes
|
||||
yes
|
||||
" | scp /etc/docker/daemon.json root@${ip}:/etc/docker/daemon.json
|
||||
|
||||
ssh root@${ip} "systemctl restart docker"
|
||||
done
|
||||
|
||||
ip_list=(10.20.1.133 10.20.1.134 10.20.1.132)
|
||||
for ip in "${ip_list[@]}"; do
|
||||
scp /etc/docker/daemon.json root@${ip}:/etc/docker/daemon.json
|
||||
# scp /etc/ssh/sshd_config root@${ip}:/etc/ssh/sshd_config
|
||||
|
||||
ssh root@${ip} "systemctl restart docker"
|
||||
# ssh root@${ip} "systemctl restart sshd"
|
||||
done
|
||||
|
||||
|
||||
vim /etc/docker/daemon.json
|
||||
{
|
||||
"insecure-registries" : ["20.47.129.116:8033"]
|
||||
}
|
||||
|
||||
systemctl restart docker
|
||||
|
||||
|
||||
list=(iptables)
|
||||
|
||||
for Packages in "${list[@]}"
|
||||
do
|
||||
apt-get download $(apt-cache depends --recurse --no-recommends --no-suggests --no-conflicts --no-breaks --no-replaces --no-enhances ${Packages} | grep "^\w" | sort -u)
|
||||
done
|
||||
|
||||
|
||||
sudo dpkg -i ./containerd.io_1.6.15-1_amd64.deb \
|
||||
./docker-ce-cli_20.10.10~3-0~ubuntu-focal_amd64.deb \
|
||||
./docker-ce_20.10.10~3-0~ubuntu-focal_amd64.deb \
|
||||
./docker-ce-rootless-extras_20.10.10~3-0~ubuntu-focal_amd64.deb \
|
||||
./docker-buildx-plugin_0.11.1-1~ubuntu.20.04~focal_amd64.deb \
|
||||
./docker-compose-plugin_2.19.1-1~ubuntu.20.04~focal_amd64.deb
|
||||
|
||||
rpcinfo -p localhost
|
||||
|
||||
# 生成ed25519 版本的ssh key
|
||||
ssh-keygen -t ed25519 -f .ssh/id_ed25519 -C "m@github"
|
||||
|
||||
echo $(cat .ssh/id_ed25519.pub)
|
||||
|
||||
echo "ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHXDzet+Z2/AmrzIZpPviz7Z9AMxLWiJUOwtj/3NPauk m@github" >>.ssh/authorized_keys
|
||||
|
||||
# 修改calico-node检测的IP
|
||||
kubectl -n kube-system edit daemonset calico-node
|
||||
env:
|
||||
- name: FELIX_INTERFACEPREFIX
|
||||
value: "eth0"
|
||||
|
||||
# 更加保险
|
||||
kubectl set env daemonset/calico-node -n kube-system IP_AUTODETECTION_METHOD=skip-interface=eth.*
|
||||
|
||||
|
||||
./mysql -uroot -pQzfXQhd3bQ -h127.0.0.1 -P33306
|
||||
|
||||
|
||||
redis-cli -h 127.0.0.1 -a Mcache@4522
|
||||
|
||||
./mysql -uboge -pboge8tingH -h127.0.0.1 -P21306
|
||||
53
23-浙江交投/1-base-env-shell/进度说明-10日版本.txt
Normal file
53
23-浙江交投/1-base-env-shell/进度说明-10日版本.txt
Normal file
@@ -0,0 +1,53 @@
|
||||
第一天
|
||||
1. 内核配置修改、优化,关闭swap、SELinux等服务器环境初始化
|
||||
2. 通用工具安装(curl wget vim ping telnet等)、配置免密登录等
|
||||
3. 安装时间同步服务器,对所有的服务器的时间进行同步
|
||||
|
||||
第二天
|
||||
1. 离线、下载并安装docker服务,修改优化docker配置
|
||||
2. 安装docker-compose 安装Harbor服务器 配置服务器免密拉取镜像
|
||||
3. 离线下载安装GlusterFS、Heketi、NFS,配置安装底层存储服务
|
||||
|
||||
第三天
|
||||
1. 离线下载Kubernetes安装文件
|
||||
2. 上传Kubernetes离线安装包到所有服务器节点
|
||||
3. 修改对应的kuberntes安装部署配置
|
||||
4. 配置apiserver、controller-manager、scheduler、kube-proxy服务文件并安装、安装master和node节点,验证Kubernetes集群安装
|
||||
|
||||
第四天
|
||||
1. 安装calico网络插件
|
||||
2. 验证Kubernetes集群,对网络、节点、镜像等验证
|
||||
3. 下载离线镜像,安装服务暴露层Ingress组件
|
||||
|
||||
第五天
|
||||
1. 下载所有中间件所需的镜像,并进行中间件的配置及安装部署
|
||||
2. 安装MySQL数据库组件,并进行配置验证
|
||||
3. 安装MongoDB并完成配置与验证
|
||||
4. 安装Redis并完成配置与验证
|
||||
5. 安装EMQX无人机飞行控制通信组件并完成配置与验证
|
||||
6. 安装Rabbit MQ消息队列组件并配置验证
|
||||
7. 安装Nacos微服务注册中心组件并配置验证
|
||||
|
||||
第六天
|
||||
1. 配置存储类、创建存储卷、安装业务层对象存储服务Minio集群
|
||||
2. 安装视频流媒体业务SRS-Cluster
|
||||
3. 确保推拉流服务正常运行
|
||||
|
||||
第七天
|
||||
1. 离线下载并上传所有的中移凌云平台业务镜像
|
||||
2. 根据环境适配中移凌云平台的业务配置并上传
|
||||
3. 安装中移凌云基础初始化数据,运营平台数据初始化
|
||||
|
||||
第八天
|
||||
1. 部署所有业务组件微服务,确保所有微服务正常运行
|
||||
2. 初步达到平台的运行(不可使用)
|
||||
|
||||
第九天
|
||||
1. 部署GDR转码服务器,完成配置信息
|
||||
2. 保证GDR服务运行正常
|
||||
3. 开始初步联调
|
||||
|
||||
第十天
|
||||
1. 中移凌云平台安装部署功能初步验证
|
||||
2. 平台业务组件运行联测
|
||||
3. 确保平台核心功能正常可用
|
||||
6
23-浙江交投/时间记录.txt
Normal file
6
23-浙江交投/时间记录.txt
Normal file
@@ -0,0 +1,6 @@
|
||||
2024-4-17 2024-4-24
|
||||
|
||||
|
||||
dltu 0 0 /root/wdd/image/rke 10.100.2.121:8033
|
||||
dltu 0 0 /root/wdd/image/middle 10.100.2.121:8033
|
||||
dltu 0 0 /root/wdd/image/rke 10.100.2.121:8033
|
||||
Reference in New Issue
Block a user