diff --git a/.fastRequest/collections/Root/Default Group/directory.json b/.fastRequest/collections/Root/Default Group/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/directory.json b/.fastRequest/collections/Root/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/ExecutionController/SyncCommandToAgentBody.rapi b/.fastRequest/collections/Root/server/ExecutionController/SyncCommandToAgentBody.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/ExecutionController/directory.json b/.fastRequest/collections/Root/server/ExecutionController/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/ExecutionController/patchCommandToAgentAll.rapi b/.fastRequest/collections/Root/server/ExecutionController/patchCommandToAgentAll.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/SchedulerController/createScriptScheduler.rapi b/.fastRequest/collections/Root/server/SchedulerController/createScriptScheduler.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/SchedulerController/directory.json b/.fastRequest/collections/Root/server/SchedulerController/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/SchedulerController/queryAllQuartzJob.rapi b/.fastRequest/collections/Root/server/SchedulerController/queryAllQuartzJob.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/SchedulerController/queryAllTriggers.rapi b/.fastRequest/collections/Root/server/SchedulerController/queryAllTriggers.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/SchedulerController/queryRunQuartzJob.rapi b/.fastRequest/collections/Root/server/SchedulerController/queryRunQuartzJob.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/StatusController/GetHealthyStatusAgentList.rapi b/.fastRequest/collections/Root/server/StatusController/GetHealthyStatusAgentList.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/StatusController/ManualUpdateAgentStatus.rapi b/.fastRequest/collections/Root/server/StatusController/ManualUpdateAgentStatus.rapi old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/StatusController/directory.json b/.fastRequest/collections/Root/server/StatusController/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/collections/Root/server/directory.json b/.fastRequest/collections/Root/server/directory.json old mode 100644 new mode 100755 diff --git a/.fastRequest/config/fastRequestCurrentProjectConfig.json b/.fastRequest/config/fastRequestCurrentProjectConfig.json old mode 100644 new mode 100755 diff --git a/.fastRequest/config/fastRequestCurrentProjectEnvironment.json b/.fastRequest/config/fastRequestCurrentProjectEnvironment.json old mode 100644 new mode 100755 diff --git a/.github/workflows/build-push-docker.yml b/.github/workflows/build-push-docker.yml old mode 100644 new mode 100755 diff --git a/.gitignore b/.gitignore old mode 100644 new mode 100755 diff --git a/.run/TestUpdateCmiiDeploymentImageTag in wdd.io_agent-operator.run.xml b/.run/TestUpdateCmiiDeploymentImageTag in wdd.io_agent-operator.run.xml old mode 100644 new mode 100755 diff --git a/.run/agent-go bastion.run.xml b/.run/agent-go bastion.run.xml old mode 100644 new mode 100755 diff --git a/.run/agent-go main.run.xml b/.run/agent-go main.run.xml old mode 100644 new mode 100755 diff --git a/.run/go build agent-go.run.xml b/.run/go build agent-go.run.xml old mode 100644 new mode 100755 diff --git a/README.md b/README.md old mode 100644 new mode 100755 diff --git a/agent-common/assert/MyAssert.go b/agent-common/assert/MyAssert.go old mode 100644 new mode 100755 diff --git a/agent-common/assert/assert.go b/agent-common/assert/assert.go old mode 100644 new mode 100755 diff --git a/agent-common/assert/assert_test.go b/agent-common/assert/assert_test.go old mode 100644 new mode 100755 diff --git a/agent-common/go.mod b/agent-common/go.mod old mode 100644 new mode 100755 diff --git a/agent-common/go.sum b/agent-common/go.sum old mode 100644 new mode 100755 diff --git a/agent-common/image/ImageNameConvert.go b/agent-common/image/ImageNameConvert.go old mode 100644 new mode 100755 diff --git a/agent-common/image/ImageNameConvert_test.go b/agent-common/image/ImageNameConvert_test.go old mode 100644 new mode 100755 diff --git a/agent-common/logger/logger.go b/agent-common/logger/logger.go old mode 100644 new mode 100755 diff --git a/agent-common/pusher/CmiiUpdateMessage.go b/agent-common/pusher/CmiiUpdateMessage.go old mode 100644 new mode 100755 diff --git a/agent-common/pusher/CmiiUpdateMessage_test.go b/agent-common/pusher/CmiiUpdateMessage_test.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/DownloadUtils.go b/agent-common/utils/DownloadUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/FileUtils.go b/agent-common/utils/FileUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/MathUtils.go b/agent-common/utils/MathUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/PrintUtils.go b/agent-common/utils/PrintUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/ReflectUtils.go b/agent-common/utils/ReflectUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/StringUtils.go b/agent-common/utils/StringUtils.go old mode 100644 new mode 100755 diff --git a/agent-common/utils/TimeUtils.go b/agent-common/utils/TimeUtils.go old mode 100644 new mode 100755 diff --git a/agent-deploy/OctopusDeploy.go b/agent-deploy/OctopusDeploy.go old mode 100644 new mode 100755 index 2735218..2526079 --- a/agent-deploy/OctopusDeploy.go +++ b/agent-deploy/OctopusDeploy.go @@ -2,6 +2,7 @@ package agent_deploy import ( "os" + "strings" image2 "wdd.io/agent-common/image" "wdd.io/agent-common/logger" "wdd.io/agent-common/utils" @@ -108,7 +109,7 @@ var IgnoreCmiiBackendAppName = map[string]string{ func backendDeploy(common *z_dep.CommonEnvironmentConfig, backendImageVersionMap map[string]string) { - os.Remove(c_app.BackendApplyFilePath) + os.Remove(z_dep.BackendApplyFilePath) for appName, tag := range backendImageVersionMap { c_app.DefaultCmiiBackendConfig.AppName = appName @@ -126,7 +127,7 @@ func backendDeploy(common *z_dep.CommonEnvironmentConfig, backendImageVersionMap func frontendDeploy(common *z_dep.CommonEnvironmentConfig, frontendImageVersionMap map[string]string) { - os.Remove(c_app.FrontendApplyFilePath) + os.Remove(z_dep.FrontendApplyFilePath) c_app.FrontendDefaultNginxDeploy(common) @@ -148,7 +149,7 @@ func frontendDeploy(common *z_dep.CommonEnvironmentConfig, frontendImageVersionM } func configMapDeploy(common *z_dep.CommonEnvironmentConfig) { - os.Remove(c_app.ConfigMapApplyFilePath) + os.Remove(z_dep.ConfigMapApplyFilePath) for frontendName, shortName := range c_app.FrontendShortNameMaps { c_app.DefaultCmiiFrontendConfig.AppName = frontendName @@ -169,25 +170,32 @@ func configMapDeploy(common *z_dep.CommonEnvironmentConfig) { // CmiiEnvironmentDeploy 部署完整的CMII环境的所有组件 func CmiiEnvironmentDeploy(isCompleteDeploy bool, cmiiNameSpace string, backendImageVersionMap, frontendImageVersionMap map[string]string) (applyYamlFolder string) { - // common environment - common := &z_dep.CommonEnvironmentConfig{ - WebIP: "lab.uavcmlc.com", - WebPort: "", - HarborIP: image2.CmiiHarborPrefix, - HarborPort: "", - Namespace: cmiiNameSpace, - TagVersion: "5.5.0", - TenantEnv: cmiiNameSpace, + folderPrefix := "/home/wdd/IdeaProjects/ProjectOctopus/agent-deploy/" + cmiiNameSpace + "/" + + tenantEnv := cmiiNameSpace + if strings.Contains(cmiiNameSpace, "-") { + split := strings.Split(cmiiNameSpace, "-") + tenantEnv = split[len(split)-1] } - folderPrefix := "/home/wdd/IdeaProjects/ProjectOctopus/agent-deploy/" + common.Namespace + "/" + // common environment + common := &z_dep.CommonEnvironmentConfig{ + WebIP: "lab.uavcmlc.com", + WebPort: "", + HarborIP: image2.CmiiHarborPrefix, + HarborPort: "", + Namespace: cmiiNameSpace, + TagVersion: "5.5.0", + TenantEnv: tenantEnv, + ApplyFilePrefix: folderPrefix, + } // clear old apply file - _ = os.Mkdir(folderPrefix, os.ModeDir) + _ = os.Mkdir(folderPrefix, os.ModePerm) oldApplyFileFolder := folderPrefix + "old" utils.RemoveFolderComplete(oldApplyFileFolder) - _ = os.Mkdir(oldApplyFileFolder, os.ModeDir) + _ = os.Mkdir(oldApplyFileFolder, os.ModePerm) // move all apply file to old folder _ = utils.FolderMoveFiles(folderPrefix, oldApplyFileFolder) @@ -196,7 +204,9 @@ func CmiiEnvironmentDeploy(isCompleteDeploy bool, cmiiNameSpace string, backendI // ignore redis rabbitmq mongo nacos emqx // sync mysql-data - z_dep.ApplyFilePrefix = folderPrefix + // generate + common.GenerateApplyFilePath() + // generate new apply file for specific environment if isCompleteDeploy { diff --git a/agent-deploy/OctopusDeploy_test.go b/agent-deploy/OctopusDeploy_test.go old mode 100644 new mode 100755 diff --git a/agent-deploy/a_dashboard/DeployK8sDashboard.go b/agent-deploy/a_dashboard/DeployK8sDashboard.go old mode 100644 new mode 100755 index ead289e..9af74bf --- a/agent-deploy/a_dashboard/DeployK8sDashboard.go +++ b/agent-deploy/a_dashboard/DeployK8sDashboard.go @@ -6,17 +6,9 @@ import ( ) var ( - K8sDashboardApplyFilePath = "" - log = logger.Log + log = logger.Log ) -func init() { - - K8sDashboardApplyFilePath = z_dep.ApplyFilePrefix + "k8s-dashboard.yaml" - - log.InfoF("K8sDashboardApplyFilePath: %s", K8sDashboardApplyFilePath) -} - func K8sDashboardDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiK8sDashboardTemplate, K8sDashboardApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiK8sDashboardTemplate, z_dep.K8sDashboardApplyFilePath) } diff --git a/agent-deploy/a_dashboard/TemplateK8SDashboard.go b/agent-deploy/a_dashboard/TemplateK8SDashboard.go old mode 100644 new mode 100755 diff --git a/agent-deploy/a_nfs/DeployNFS.go b/agent-deploy/a_nfs/DeployNFS.go old mode 100644 new mode 100755 index dd432c5..c1e1f51 --- a/agent-deploy/a_nfs/DeployNFS.go +++ b/agent-deploy/a_nfs/DeployNFS.go @@ -6,28 +6,17 @@ import ( ) var ( - NfsApplyFilePath = "" - NfsTestApplyFilePath = "" - log = logger.Log + log = logger.Log ) type NfsDeployConfig struct { NfsLocalPath string } -func init() { - - NfsApplyFilePath = z_dep.ApplyFilePrefix + "k8s-nfs.yaml" - NfsTestApplyFilePath = z_dep.ApplyFilePrefix + "k8s-nfs-test.yaml" - - log.InfoF("NfsApplyFilePath : %s\n", NfsApplyFilePath) - log.InfoF("NfsTestApplyFilePath : %s\n", NfsTestApplyFilePath) -} - func NFSDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiNfsTemplate, NfsApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiNfsTemplate, z_dep.NfsApplyFilePath) } func NFSTestDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiNFSTestTemplate, NfsTestApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiNFSTestTemplate, z_dep.NfsTestApplyFilePath) } diff --git a/agent-deploy/a_nfs/TemplateNFS.go b/agent-deploy/a_nfs/TemplateNFS.go old mode 100644 new mode 100755 diff --git a/agent-deploy/a_nfs/TemplateNFSTest.go b/agent-deploy/a_nfs/TemplateNFSTest.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiEmqxTemplate.go b/agent-deploy/b_middle/CmiiEmqxTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiMongoTemplate.go b/agent-deploy/b_middle/CmiiMongoTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiMySQLTemplate.go b/agent-deploy/b_middle/CmiiMySQLTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiNacosTemplate.go b/agent-deploy/b_middle/CmiiNacosTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiRabbitMQTemplate.go b/agent-deploy/b_middle/CmiiRabbitMQTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/CmiiRedisTemplate.go b/agent-deploy/b_middle/CmiiRedisTemplate.go old mode 100644 new mode 100755 diff --git a/agent-deploy/b_middle/DeployMiddleware.go b/agent-deploy/b_middle/DeployMiddleware.go old mode 100644 new mode 100755 index 2e68d05..a280845 --- a/agent-deploy/b_middle/DeployMiddleware.go +++ b/agent-deploy/b_middle/DeployMiddleware.go @@ -6,57 +6,31 @@ import ( ) var ( - EmqxApplyFilePath = "" - MongoApplyFilePath = "" - RabbitMQApplyFilePath = "" - RedisApplyFilePath = "" - MySQLApplyFilePath = "" - NacosApplyFilePath = "" - PVCApplyFilePath = "" - log = logger.Log + log = logger.Log ) -func init() { - - EmqxApplyFilePath = z_dep.ApplyFilePrefix + "k8s-emqx.yaml" - MongoApplyFilePath = z_dep.ApplyFilePrefix + "k8s-mongo.yaml" - RabbitMQApplyFilePath = z_dep.ApplyFilePrefix + "k8s-rabbitmq.yaml" - RedisApplyFilePath = z_dep.ApplyFilePrefix + "k8s-redis.yaml" - MySQLApplyFilePath = z_dep.ApplyFilePrefix + "k8s-mysql.yaml" - NacosApplyFilePath = z_dep.ApplyFilePrefix + "k8s-nacos.yaml" - PVCApplyFilePath = z_dep.ApplyFilePrefix + "k8s-pvc.yaml" - - log.DebugF("EmqxApplyFilePath: %s", EmqxApplyFilePath) - log.DebugF("MongoApplyFilePath: %s", MongoApplyFilePath) - log.DebugF("RabbitMQApplyFilePath: %s", RabbitMQApplyFilePath) - log.DebugF("RedisApplyFilePath: %s", RedisApplyFilePath) - log.DebugF("MySQLApplyFilePath: %s", MySQLApplyFilePath) - log.DebugF("NacosApplyFilePath: %s", NacosApplyFilePath) - log.DebugF("PVCApplyFilePath: %s", PVCApplyFilePath) -} - func MidEmqxDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiEmqxTemplate, EmqxApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiEmqxTemplate, z_dep.EmqxApplyFilePath) } func MidMongoDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiMongoTemplate, MongoApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiMongoTemplate, z_dep.MongoApplyFilePath) } func MidRabbitMQDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiRabbitMQTemplate, RabbitMQApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiRabbitMQTemplate, z_dep.RabbitMQApplyFilePath) } func MidRedisDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiRedisTemplate, RedisApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiRedisTemplate, z_dep.RedisApplyFilePath) } func MidMySQlDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiMySQLTemplate, MySQLApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiMySQLTemplate, z_dep.MySQLApplyFilePath) } func MidNacosDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiNacosTemplate, NacosApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiNacosTemplate, z_dep.NacosApplyFilePath) } func PVCDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiPVCTemplate, PVCApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiPVCTemplate, z_dep.PVCApplyFilePath) } diff --git a/agent-deploy/b_middle/TemplateCmiiPVC.go b/agent-deploy/b_middle/TemplateCmiiPVC.go old mode 100644 new mode 100755 diff --git a/agent-deploy/c_app/DeployCmiiApp.go b/agent-deploy/c_app/DeployCmiiApp.go old mode 100644 new mode 100755 index 56614e3..7b67638 --- a/agent-deploy/c_app/DeployCmiiApp.go +++ b/agent-deploy/c_app/DeployCmiiApp.go @@ -32,27 +32,8 @@ type CmiiFrontendConfig struct { var ( DefaultCmiiBackendConfig = &CmiiBackendConfig{} DefaultCmiiFrontendConfig = &CmiiFrontendConfig{} - BackendApplyFilePath = "" - FrontendApplyFilePath = "" - SRSApplyFilePath = "" - IngresApplyFilePath = "" - ConfigMapApplyFilePath = "" ) -func init() { - BackendApplyFilePath = z_dep.ApplyFilePrefix + "k8s-backend.yaml" - FrontendApplyFilePath = z_dep.ApplyFilePrefix + "k8s-frontend.yaml" - SRSApplyFilePath = z_dep.ApplyFilePrefix + "k8s-srs.yaml" - IngresApplyFilePath = z_dep.ApplyFilePrefix + "k8s-ingress.yaml" - ConfigMapApplyFilePath = z_dep.ApplyFilePrefix + "k8s-configmap.yaml" - - log.DebugF("backend apply file path: %s\n", BackendApplyFilePath) - log.DebugF("frontend apply file path: %s\n", FrontendApplyFilePath) - log.DebugF("srs apply file path: %s\n", SRSApplyFilePath) - log.DebugF("ingress apply file path: %s\n", IngresApplyFilePath) - log.DebugF("config map apply file path: %s\n", ConfigMapApplyFilePath) -} - func (backend *CmiiBackendConfig) BackendDeploy(common *z_dep.CommonEnvironmentConfig) bool { // copy @@ -65,16 +46,16 @@ func (backend *CmiiBackendConfig) BackendDeploy(common *z_dep.CommonEnvironmentC return false } - if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendDeploymentTemplate, BackendApplyFilePath) { + if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendDeploymentTemplate, z_dep.BackendApplyFilePath) { return false } - if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendServiceTemplate, BackendApplyFilePath) { + if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendServiceTemplate, z_dep.BackendApplyFilePath) { return false } // pvc if backend.NeedPvcCache { - if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendPVCTemplate, BackendApplyFilePath) { + if !z_dep.ParseEnvToApplyFile(backend, CmiiBackendPVCTemplate, z_dep.BackendApplyFilePath) { return false } } @@ -94,10 +75,10 @@ func (frontend *CmiiFrontendConfig) FrontendDeploy(common *z_dep.CommonEnvironme return false } - if !z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendDeploymentTemplate, FrontendApplyFilePath) { + if !z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendDeploymentTemplate, z_dep.FrontendApplyFilePath) { return false } - if !z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendServiceTemplate, FrontendApplyFilePath) { + if !z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendServiceTemplate, z_dep.FrontendApplyFilePath) { return false } @@ -115,19 +96,19 @@ func (frontend *CmiiFrontendConfig) ConfigMapDeploy(commonEnv *z_dep.CommonEnvir return false } - return z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendConfigMapTemplate, ConfigMapApplyFilePath) + return z_dep.ParseEnvToApplyFile(frontend, CmiiFrontendConfigMapTemplate, z_dep.ConfigMapApplyFilePath) } func IngressDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - if !commonEnv.ParseCommonEnvToApplyFile(CmiiFrontendIngressTemplate, IngresApplyFilePath) { + if !commonEnv.ParseCommonEnvToApplyFile(CmiiFrontendIngressTemplate, z_dep.IngresApplyFilePath) { return false } - if !commonEnv.ParseCommonEnvToApplyFile(CmiiBackendIngressTemplate, IngresApplyFilePath) { + if !commonEnv.ParseCommonEnvToApplyFile(CmiiBackendIngressTemplate, z_dep.IngresApplyFilePath) { return false } - if !commonEnv.ParseCommonEnvToApplyFile(CmiiGatewayIngressTemplate, IngresApplyFilePath) { + if !commonEnv.ParseCommonEnvToApplyFile(CmiiGatewayIngressTemplate, z_dep.IngresApplyFilePath) { return false } @@ -135,10 +116,10 @@ func IngressDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { } func SRSDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - os.Remove(SRSApplyFilePath) - return commonEnv.ParseCommonEnvToApplyFile(CmiiSrsTemplate, SRSApplyFilePath) + os.Remove(z_dep.SRSApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiSrsTemplate, z_dep.SRSApplyFilePath) } func FrontendDefaultNginxDeploy(commonEnv *z_dep.CommonEnvironmentConfig) bool { - return commonEnv.ParseCommonEnvToApplyFile(CmiiFrontendDefaultNginxConfTemplate, FrontendApplyFilePath) + return commonEnv.ParseCommonEnvToApplyFile(CmiiFrontendDefaultNginxConfTemplate, z_dep.FrontendApplyFilePath) } diff --git a/agent-deploy/c_app/DeployCmiiApp_test.go b/agent-deploy/c_app/DeployCmiiApp_test.go old mode 100644 new mode 100755 diff --git a/agent-deploy/c_app/FrontendConfigMap.go b/agent-deploy/c_app/FrontendConfigMap.go old mode 100644 new mode 100755 index abbc70d..7a6602b --- a/agent-deploy/c_app/FrontendConfigMap.go +++ b/agent-deploy/c_app/FrontendConfigMap.go @@ -25,6 +25,9 @@ var FrontendShortNameMaps = map[string]string{ "cmii-uav-platform-threedsimulation": "threedsimulation", "cmii-uav-platform-jiangsuwenlv": "jiangsuwenlv", "cmii-uav-platform-qinghaitourism": "qinghaitourism", + "cmii-uav-platform-qingdao": "qingdao", + "cmii-uav-platform-hljtt": "hljtt", + "cmii-uav-platform-visualization": "visualization", } var FrontendClientIdMaps = map[string]string{ @@ -54,4 +57,5 @@ var FrontendClientIdMaps = map[string]string{ "cmii-uav-platform-visualization": "empty", "cmii-uav-platform-traffic": "APP_Jc8i2wOQ1t73QEJS", "cmii-uav-platform-jiangsuwenlv": "empty", + "cmii-uav-platform-hljtt": "empty", } diff --git a/agent-deploy/c_app/TemplateCmiiBackend.go b/agent-deploy/c_app/TemplateCmiiBackend.go old mode 100644 new mode 100755 diff --git a/agent-deploy/c_app/TemplateCmiiFrontend.go b/agent-deploy/c_app/TemplateCmiiFrontend.go old mode 100644 new mode 100755 diff --git a/agent-deploy/c_app/TemplateCmiiSRS.go b/agent-deploy/c_app/TemplateCmiiSRS.go old mode 100644 new mode 100755 diff --git a/agent-deploy/c_app/TemplateIngressConfigMap.go b/agent-deploy/c_app/TemplateIngressConfigMap.go old mode 100644 new mode 100755 diff --git a/agent-deploy/go.mod b/agent-deploy/go.mod old mode 100644 new mode 100755 diff --git a/agent-deploy/go.sum b/agent-deploy/go.sum old mode 100644 new mode 100755 diff --git a/agent-deploy/uavcloud-dev/k8s-backend.yaml b/agent-deploy/uavcloud-dev/k8s-backend.yaml new file mode 100644 index 0000000..6dd03b4 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-backend.yaml @@ -0,0 +1,5282 @@ +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-multilink + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-multilink + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-multilink + image: harbor.cdcyy.com.cn/cmii/cmii-uav-multilink:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-multilink + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-multilink + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-multilink + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-multilink + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-autowaypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-autowaypoint + image: harbor.cdcyy.com.cn/cmii/cmii-uav-autowaypoint:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-autowaypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-autowaypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-autowaypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-brain + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-brain + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-brain + image: harbor.cdcyy.com.cn/cmii/cmii-uav-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-brain + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-brain + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-brain + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-brain + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-device + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-device + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-device + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-device + image: harbor.cdcyy.com.cn/cmii/cmii-uav-device:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-device + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-device + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-device + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-device + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uas-lifecycle + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uas-lifecycle + image: harbor.cdcyy.com.cn/cmii/cmii-uas-lifecycle:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uas-lifecycle + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uas-lifecycle + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uas-lifecycle + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-industrial-portfolio + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-industrial-portfolio + image: harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-industrial-portfolio + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-industrial-portfolio + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-industrial-portfolio + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-uav-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-developer + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-developer + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-developer + image: harbor.cdcyy.com.cn/cmii/cmii-uav-developer:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-developer + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-developer + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-developer + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-developer + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-integration + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-integration + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-integration + image: harbor.cdcyy.com.cn/cmii/cmii-uav-integration:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-integration + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-integration + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-integration + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-integration + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cloud-live + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cloud-live + image: harbor.cdcyy.com.cn/cmii/cmii-uav-cloud-live:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-cloud-live + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-cloud-live + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cloud-live + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-app-release + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-app-release + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-app-release + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-app-release + image: harbor.cdcyy.com.cn/cmii/cmii-app-release:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-app-release + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-app-release + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-app-release + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-app-release + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-user + image: harbor.cdcyy.com.cn/cmii/cmii-uav-user:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-data-post-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-data-post-process + image: harbor.cdcyy.com.cn/cmii/cmii-uav-data-post-process:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-data-post-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-data-post-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-data-post-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-user + image: harbor.cdcyy.com.cn/cmii/cmii-admin-user:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-alarm + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-alarm + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-alarm + image: harbor.cdcyy.com.cn/cmii/cmii-uav-alarm:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-alarm + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-alarm + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-alarm + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-alarm + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-emergency + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-emergency + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-emergency + image: harbor.cdcyy.com.cn/cmii/cmii-uav-emergency:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-emergency + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-emergency + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-emergency + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-emergency + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-oauth + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-oauth + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-oauth + image: harbor.cdcyy.com.cn/cmii/cmii-uav-oauth:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-oauth + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-oauth + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-oauth + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-oauth + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-supervision + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-suav-supervision + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-supervision + image: harbor.cdcyy.com.cn/cmii/cmii-suav-supervision:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-supervision + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-suav-supervision + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-supervision + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-suav-supervision + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-process + image: harbor.cdcyy.com.cn/cmii/cmii-uav-process:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-admin-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cms + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cms + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cms + image: harbor.cdcyy.com.cn/cmii/cmii-uav-cms:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-cms + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-cms + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cms + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cms + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-clusters + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-clusters + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-clusters + image: harbor.cdcyy.com.cn/cmii/cmii-uav-clusters:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-clusters + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-clusters + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-clusters + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-clusters + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-threedsimulation + image: harbor.cdcyy.com.cn/cmii/cmii-uav-threedsimulation:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-threedsimulation + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-threedsimulation + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-depotautoreturn + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-depotautoreturn + image: harbor.cdcyy.com.cn/cmii/cmii-uav-depotautoreturn:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-depotautoreturn + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-depotautoreturn + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-depotautoreturn + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-airspace + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-airspace + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-airspace + image: harbor.cdcyy.com.cn/cmii/cmii-uav-airspace:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-airspace + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-airspace + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-airspace + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-airspace + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-waypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-waypoint + image: harbor.cdcyy.com.cn/cmii/cmii-uav-waypoint:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-waypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-waypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-waypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-waypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-open-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-open-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-open-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-open-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-open-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-open-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-open-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-open-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mission + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mission + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mission + image: harbor.cdcyy.com.cn/cmii/cmii-uav-mission:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-mission + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-mission + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mission + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mission + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-notice + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-notice + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-notice + image: harbor.cdcyy.com.cn/cmii/cmii-uav-notice:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-notice + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-notice + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-notice + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-notice + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-logger + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-logger + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-logger + image: harbor.cdcyy.com.cn/cmii/cmii-uav-logger:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-logger + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-logger + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-logger + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-logger + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mqtthandler + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mqtthandler + image: harbor.cdcyy.com.cn/cmii/cmii-uav-mqtthandler:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-mqtthandler + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-mqtthandler + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mqtthandler + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uas-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uas-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uas-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-uas-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uas-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uas-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uas-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uas-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-kpi-monitor + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-kpi-monitor + image: harbor.cdcyy.com.cn/cmii/cmii-uav-kpi-monitor:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-kpi-monitor + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-kpi-monitor + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-kpi-monitor + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-surveillance + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-surveillance + image: harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-surveillance + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-surveillance + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-surveillance + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-surveillance + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-bridge + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-bridge + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-bridge + image: harbor.cdcyy.com.cn/cmii/cmii-uav-bridge:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-bridge + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-bridge + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-bridge + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-bridge + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-data + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-data + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-data + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-data + image: harbor.cdcyy.com.cn/cmii/cmii-admin-data:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-data + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-data + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-data + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-data + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-material-warehouse + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-material-warehouse + image: harbor.cdcyy.com.cn/cmii/cmii-uav-material-warehouse:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-material-warehouse + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-material-warehouse + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-material-warehouse + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-tower + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-tower + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-tower + image: harbor.cdcyy.com.cn/cmii/cmii-uav-tower:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-tower + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-tower + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-tower + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-tower + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 diff --git a/agent-deploy/uavcloud-dev/k8s-configmap.yaml b/agent-deploy/uavcloud-dev/k8s-configmap.yaml new file mode 100644 index 0000000..9b2669e --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-configmap.yaml @@ -0,0 +1,378 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-open + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "open", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-security + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "security", + AppClientId: "APP_JUSEMc7afyWXxvE7" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-securityh5 + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "securityh5", + AppClientId: "APP_N3ImO0Ubfu9peRHD" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-hljtt + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "hljtt", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-base + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "base", + AppClientId: "APP_9LY41OaKSqk2btY0" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-detection + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "detection", + AppClientId: "APP_FDHW2VLVDWPnnOCy" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-multiterminal + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "multiterminal", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-mws + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "mws", + AppClientId: "APP_uKniXPELlRERBBwK" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-splice + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "splice", + AppClientId: "APP_zE0M3sTRXrCIJS8Y" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-traffic + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "traffic", + AppClientId: "APP_Jc8i2wOQ1t73QEJS" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-qinghaitourism + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "qinghaitourism", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-media + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "media", + AppClientId: "APP_4AU8lbifESQO4FD6" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-threedsimulation + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "threedsimulation", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-jiangsuwenlv + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "jiangsuwenlv", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-visualization + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "visualization", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-armypeople + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "armypeople", + AppClientId: "APP_UIegse6Lfou9pO1U" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-ai-brain + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "ai-brain", + AppClientId: "APP_rafnuCAmBESIVYMH" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-oms + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "oms", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-qingdao + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "qingdao", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervision + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "supervision", + AppClientId: "APP_qqSu82THfexI8PLM" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervisionh5 + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "supervisionh5", + AppClientId: "APP_qqSu82THfexI8PLM" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-pangu + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-logistics + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "logistics", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-seniclive + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "seniclive", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-share + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "share", + AppClientId: "APP_4lVSVI0ZGxTssir8" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-cmsportal + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "cmsportal", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-emergency + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "emergency", + AppClientId: "APP_aGsTAY1uMZrpKdfk" + } diff --git a/agent-deploy/uavcloud-dev/k8s-emqx.yaml b/agent-deploy/uavcloud-dev/k8s-emqx.yaml new file mode 100644 index 0000000..e5d1176 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-emqx.yaml @@ -0,0 +1,263 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: uavcloud-dev +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: uavcloud-dev + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: harbor.cdcyy.com.cn/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: uavcloud-dev +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: uavcloud-dev +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: uavcloud-dev +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 diff --git a/agent-deploy/uavcloud-dev/k8s-frontend.yaml b/agent-deploy/uavcloud-dev/k8s-frontend.yaml new file mode 100644 index 0000000..d5f997a --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-frontend.yaml @@ -0,0 +1,2354 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: nginx-cm + namespace: uavcloud-dev + labels: + cmii.type: frontend +data: + nginx.conf: | + user root; + worker_processes auto; + + events { + worker_connections 1024; + } + http { + include mime.types; + default_type application/octet-stream; + + sendfile on; + + keepalive_timeout 600; + + server { + listen 9528; + server_name localhost; + gzip on; + + location / { + root /home/cmii-platform/dist; + index index.html index.htm; + } + + error_page 500 502 503 504 /50x.html; + location = /50x.html { + root html; + } + } + } +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-pangu + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-open + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-open + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-open:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-open + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-open + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-open + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-multiterminal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-multiterminal + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-multiterminal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-multiterminal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-multiterminal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-multiterminal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-oms + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-oms + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-oms:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-oms + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-oms + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-oms + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-qingdao + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-qingdao + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-qingdao:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-qingdao + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-qingdao + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-qingdao + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-emergency-rescue + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-emergency-rescue + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-emergency-rescue:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-emergency-rescue + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-emergency + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-emergency-rescue + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-base + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-base + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-base:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-base + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-base + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-base + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-visualization + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-visualization + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-visualization:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-visualization + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-visualization + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-visualization + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-detection + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-detection + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-detection:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-detection + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-detection + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-detection + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-splice + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-splice + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-splice:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-splice + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-splice + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-splice + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-logistics + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-logistics + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-logistics:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-logistics + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-logistics + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-logistics + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervision + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervision + image: harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervision:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-platform-supervision + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervision + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervision + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-mws + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-mws + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-mws:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-mws + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-mws + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-mws + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-hljtt + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-hljtt + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-hljtt:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-hljtt + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-hljtt + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-hljtt + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-qinghaitourism + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-qinghaitourism:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-qinghaitourism + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-qinghaitourism + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-seniclive + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-seniclive + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-seniclive + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-seniclive + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-seniclive + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-securityh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-securityh5 + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-securityh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-securityh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-securityh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-securityh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-media + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-media + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-media:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-media + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-media + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-media + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervisionh5 + image: harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervisionh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-platform-supervisionh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervisionh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-jiangsuwenlv + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-jiangsuwenlv:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-jiangsuwenlv + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-jiangsuwenlv + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-armypeople + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-armypeople + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-armypeople:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-armypeople + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-armypeople + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-armypeople + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-threedsimulation + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-threedsimulation:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-threedsimulation + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-threedsimulation + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-share + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-share + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-share:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-share + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-share + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-share + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-ai-brain + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-ai-brain + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-ai-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-ai-brain + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-ai-brain + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-ai-brain + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-cms-portal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-cms-portal + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-cms-portal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-cms-portal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-cmsportal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-cms-portal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-security + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-security + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-security:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-security + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-security + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-security + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 diff --git a/agent-deploy/uavcloud-dev/k8s-ingress.yaml b/agent-deploy/uavcloud-dev/k8s-ingress.yaml new file mode 100644 index 0000000..21dc54e --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-ingress.yaml @@ -0,0 +1,544 @@ +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: uavcloud-dev + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.uavcloud-dev.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: uavcloud-dev + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: uavcloud-dev + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.uavcloud-dev.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 diff --git a/agent-deploy/uavcloud-dev/k8s-mongo.yaml b/agent-deploy/uavcloud-dev/k8s-mongo.yaml new file mode 100644 index 0000000..a40f2a1 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-mongo.yaml @@ -0,0 +1,75 @@ +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: harbor.cdcyy.com.cn/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- diff --git a/agent-deploy/uavcloud-dev/k8s-mysql.yaml b/agent-deploy/uavcloud-dev/k8s-mysql.yaml new file mode 100644 index 0000000..1834d25 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-mysql.yaml @@ -0,0 +1,421 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: harbor.cdcyy.com.cn/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv diff --git a/agent-deploy/uavcloud-dev/k8s-nacos.yaml b/agent-deploy/uavcloud-dev/k8s-nacos.yaml new file mode 100644 index 0000000..6326e2a --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-nacos.yaml @@ -0,0 +1,130 @@ +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: harbor.cdcyy.com.cn/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + - containerPort: 9848 + name: tcp-9848 + - containerPort: 9849 + name: tcp-9849 + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- diff --git a/agent-deploy/uavcloud-dev/k8s-pvc.yaml b/agent-deploy/uavcloud-dev/k8s-pvc.yaml new file mode 100644 index 0000000..cce0138 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-pvc.yaml @@ -0,0 +1,76 @@ +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi diff --git a/agent-deploy/uavcloud-dev/k8s-rabbitmq.yaml b/agent-deploy/uavcloud-dev/k8s-rabbitmq.yaml new file mode 100644 index 0000000..75ac663 --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-rabbitmq.yaml @@ -0,0 +1,327 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: harbor.cdcyy.com.cn/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq diff --git a/agent-deploy/uavcloud-dev/k8s-redis.yaml b/agent-deploy/uavcloud-dev/k8s-redis.yaml new file mode 100644 index 0000000..5d8461c --- /dev/null +++ b/agent-deploy/uavcloud-dev/k8s-redis.yaml @@ -0,0 +1,582 @@ +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: harbor.cdcyy.com.cn/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.uavcloud-dev.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + diff --git a/agent-deploy/uavcloud-dev/old/k8s-backend.yaml b/agent-deploy/uavcloud-dev/old/k8s-backend.yaml new file mode 100644 index 0000000..05d2ab0 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-backend.yaml @@ -0,0 +1,5282 @@ +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-clusters + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-clusters + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-clusters + image: harbor.cdcyy.com.cn/cmii/cmii-uav-clusters:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-clusters + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-clusters + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-clusters + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-clusters + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-clusters + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-bridge + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-bridge + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-bridge + image: harbor.cdcyy.com.cn/cmii/cmii-uav-bridge:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-bridge + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-bridge + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-bridge + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-bridge + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-bridge + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-app-release + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-app-release + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-app-release + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-app-release + image: harbor.cdcyy.com.cn/cmii/cmii-app-release:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-app-release + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-app-release + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-app-release + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-app-release + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mqtthandler + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mqtthandler + image: harbor.cdcyy.com.cn/cmii/cmii-uav-mqtthandler:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-mqtthandler + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-mqtthandler + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mqtthandler + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-uav-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mission + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mission + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mission + image: harbor.cdcyy.com.cn/cmii/cmii-uav-mission:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-mission + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-mission + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mission + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mission + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cms + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cms + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cms + image: harbor.cdcyy.com.cn/cmii/cmii-uav-cms:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-cms + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-cms + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cms + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cms + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-brain + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-brain + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-brain + image: harbor.cdcyy.com.cn/cmii/cmii-uav-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-brain + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-brain + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-brain + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-brain + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-emergency + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-emergency + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-emergency + image: harbor.cdcyy.com.cn/cmii/cmii-uav-emergency:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-emergency + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-emergency + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-emergency + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-emergency + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-industrial-portfolio + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-industrial-portfolio + image: harbor.cdcyy.com.cn/cmii/cmii-uav-industrial-portfolio:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-industrial-portfolio + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-industrial-portfolio + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-industrial-portfolio + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-threedsimulation + image: harbor.cdcyy.com.cn/cmii/cmii-uav-threedsimulation:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-threedsimulation + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-threedsimulation + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-integration + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-integration + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-integration + image: harbor.cdcyy.com.cn/cmii/cmii-uav-integration:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-integration + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-integration + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-integration + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-integration + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-data-post-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-data-post-process + image: harbor.cdcyy.com.cn/cmii/cmii-uav-data-post-process:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-data-post-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-data-post-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-data-post-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cloud-live + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cloud-live + image: harbor.cdcyy.com.cn/cmii/cmii-uav-cloud-live:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-cloud-live + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-cloud-live + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cloud-live + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-user + image: harbor.cdcyy.com.cn/cmii/cmii-uav-user:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-admin-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uas-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uas-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uas-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-uas-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uas-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uas-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uas-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uas-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-data + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-data + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-data + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-data + image: harbor.cdcyy.com.cn/cmii/cmii-admin-data:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-data + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-data + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-data + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-data + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-process + image: harbor.cdcyy.com.cn/cmii/cmii-uav-process:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-process + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-logger + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-logger + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-logger + image: harbor.cdcyy.com.cn/cmii/cmii-uav-logger:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-logger + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-logger + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-logger + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-logger + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-supervision + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-suav-supervision + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-supervision + image: harbor.cdcyy.com.cn/cmii/cmii-suav-supervision:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-supervision + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-suav-supervision + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-supervision + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-suav-supervision + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-oauth + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-oauth + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-oauth + image: harbor.cdcyy.com.cn/cmii/cmii-uav-oauth:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-oauth + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-oauth + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-oauth + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-oauth + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-tower + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-tower + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-tower + image: harbor.cdcyy.com.cn/cmii/cmii-uav-tower:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-tower + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-tower + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-tower + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-tower + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-airspace + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-airspace + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-airspace + image: harbor.cdcyy.com.cn/cmii/cmii-uav-airspace:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-airspace + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-airspace + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-airspace + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-airspace + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-open-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-open-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-open-gateway + image: harbor.cdcyy.com.cn/cmii/cmii-open-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-open-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-open-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-open-gateway + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-open-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-surveillance + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-surveillance + image: harbor.cdcyy.com.cn/cmii/cmii-uav-surveillance:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-surveillance + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-surveillance + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-surveillance + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-surveillance + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uas-lifecycle + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uas-lifecycle + image: harbor.cdcyy.com.cn/cmii/cmii-uas-lifecycle:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uas-lifecycle + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uas-lifecycle + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uas-lifecycle + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uas-lifecycle + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-device + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-device + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-device + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-device + image: harbor.cdcyy.com.cn/cmii/cmii-uav-device:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-device + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-device + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-device + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-device + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-notice + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-notice + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-notice + image: harbor.cdcyy.com.cn/cmii/cmii-uav-notice:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-notice + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-notice + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-notice + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-notice + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-waypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-waypoint + image: harbor.cdcyy.com.cn/cmii/cmii-uav-waypoint:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-waypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-waypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-waypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-waypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-user + image: harbor.cdcyy.com.cn/cmii/cmii-admin-user:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-admin-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-admin-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-user + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-developer + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-developer + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-developer + image: harbor.cdcyy.com.cn/cmii/cmii-uav-developer:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-developer + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-developer + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-developer + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-developer + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-kpi-monitor + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-kpi-monitor + image: harbor.cdcyy.com.cn/cmii/cmii-uav-kpi-monitor:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-kpi-monitor + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-kpi-monitor + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-kpi-monitor + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-autowaypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-autowaypoint + image: harbor.cdcyy.com.cn/cmii/cmii-uav-autowaypoint:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-autowaypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-autowaypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-autowaypoint + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-depotautoreturn + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-depotautoreturn + image: harbor.cdcyy.com.cn/cmii/cmii-uav-depotautoreturn:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-depotautoreturn + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-depotautoreturn + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-depotautoreturn + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-depotautoreturn + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-alarm + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-alarm + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-alarm + image: harbor.cdcyy.com.cn/cmii/cmii-uav-alarm:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-alarm + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-alarm + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-alarm + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-alarm + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-material-warehouse + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-material-warehouse + image: harbor.cdcyy.com.cn/cmii/cmii-uav-material-warehouse:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-material-warehouse + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-material-warehouse + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-material-warehouse + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-multilink + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 0 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-multilink + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-multilink + image: harbor.cdcyy.com.cn/cmii/cmii-uav-multilink:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-multilink + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: uavcloud-dev/cmii-uav-multilink + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-multilink + namespace: uavcloud-dev + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-multilink + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 diff --git a/agent-deploy/uavcloud-dev/old/k8s-configmap.yaml b/agent-deploy/uavcloud-dev/old/k8s-configmap.yaml new file mode 100644 index 0000000..3fc13e8 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-configmap.yaml @@ -0,0 +1,378 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-security + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "security", + AppClientId: "APP_JUSEMc7afyWXxvE7" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-qinghaitourism + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "qinghaitourism", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-hljtt + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "hljtt", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervision + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "supervision", + AppClientId: "APP_qqSu82THfexI8PLM" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-open + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "open", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-media + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "media", + AppClientId: "APP_4AU8lbifESQO4FD6" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-multiterminal + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "multiterminal", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-oms + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "oms", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-share + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "share", + AppClientId: "APP_4lVSVI0ZGxTssir8" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-pangu + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-base + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "base", + AppClientId: "APP_9LY41OaKSqk2btY0" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-ai-brain + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "ai-brain", + AppClientId: "APP_rafnuCAmBESIVYMH" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-mws + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "mws", + AppClientId: "APP_uKniXPELlRERBBwK" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervisionh5 + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "supervisionh5", + AppClientId: "APP_qqSu82THfexI8PLM" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-threedsimulation + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "threedsimulation", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-cmsportal + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "cmsportal", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-seniclive + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "seniclive", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-emergency + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "emergency", + AppClientId: "APP_aGsTAY1uMZrpKdfk" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-jiangsuwenlv + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "jiangsuwenlv", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-detection + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "detection", + AppClientId: "APP_FDHW2VLVDWPnnOCy" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-logistics + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "logistics", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-splice + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "splice", + AppClientId: "APP_zE0M3sTRXrCIJS8Y" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-traffic + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "traffic", + AppClientId: "APP_Jc8i2wOQ1t73QEJS" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-qingdao + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "qingdao", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-visualization + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "visualization", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-armypeople + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "armypeople", + AppClientId: "APP_UIegse6Lfou9pO1U" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-securityh5 + namespace: uavcloud-dev +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "dev", + CloudHOST: "lab.uavcmlc.com", + ApplicationShortName: "securityh5", + AppClientId: "APP_N3ImO0Ubfu9peRHD" + } diff --git a/agent-deploy/uavcloud-dev/old/k8s-emqx.yaml b/agent-deploy/uavcloud-dev/old/k8s-emqx.yaml new file mode 100644 index 0000000..e5d1176 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-emqx.yaml @@ -0,0 +1,263 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: uavcloud-dev +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: uavcloud-dev + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: harbor.cdcyy.com.cn/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: uavcloud-dev +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: uavcloud-dev +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: uavcloud-dev +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: uavcloud-dev + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 diff --git a/agent-deploy/uavcloud-dev/old/k8s-frontend.yaml b/agent-deploy/uavcloud-dev/old/k8s-frontend.yaml new file mode 100644 index 0000000..0ef87ac --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-frontend.yaml @@ -0,0 +1,2354 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: nginx-cm + namespace: uavcloud-dev + labels: + cmii.type: frontend +data: + nginx.conf: | + user root; + worker_processes auto; + + events { + worker_connections 1024; + } + http { + include mime.types; + default_type application/octet-stream; + + sendfile on; + + keepalive_timeout 600; + + server { + listen 9528; + server_name localhost; + gzip on; + + location / { + root /home/cmii-platform/dist; + index index.html index.htm; + } + + error_page 500 502 503 504 /50x.html; + location = /50x.html { + root html; + } + } + } +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-seniclive + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-seniclive + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-seniclive:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-seniclive + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-seniclive + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-seniclive + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-seniclive + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-visualization + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-visualization + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-visualization:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-visualization + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-visualization + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-visualization + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-visualization + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-hljtt + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-hljtt + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-hljtt:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-hljtt + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-hljtt + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-hljtt + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-hljtt + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-qinghaitourism + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-qinghaitourism:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-qinghaitourism + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-qinghaitourism + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-security + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-security + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-security:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-security + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-security + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-security + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervision + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervision + image: harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervision:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-platform-supervision + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervision + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervision + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-jiangsuwenlv + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-jiangsuwenlv:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-jiangsuwenlv + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-jiangsuwenlv + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-media + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-media + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-media:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-media + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-media + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-media + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-open + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-open + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-open:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-open + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-open + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-open + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-emergency-rescue + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-emergency-rescue + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-emergency-rescue:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-emergency-rescue + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-emergency + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-emergency-rescue + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-emergency-rescue + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-securityh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-securityh5 + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-securityh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-securityh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-securityh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-securityh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-detection + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-detection + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-detection:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-detection + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-detection + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-detection + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-cms-portal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-cms-portal + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-cms-portal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-cms-portal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-cmsportal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-cms-portal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-multiterminal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-multiterminal + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-multiterminal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-multiterminal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-multiterminal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-multiterminal + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-mws + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-mws + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-mws:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-mws + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-mws + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-mws + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-qingdao + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-qingdao + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-qingdao:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-qingdao + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-qingdao + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-qingdao + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-qingdao + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-ai-brain + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-ai-brain + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-ai-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-ai-brain + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-ai-brain + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-ai-brain + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-armypeople + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-armypeople + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-armypeople:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-armypeople + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-armypeople + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-armypeople + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-base + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-base + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-base:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-base + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-base + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-base + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-threedsimulation + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-threedsimulation:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-threedsimulation + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-threedsimulation + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-threedsimulation + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-logistics + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-logistics + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-logistics:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-logistics + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-logistics + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-logistics + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervisionh5 + image: harbor.cdcyy.com.cn/cmii/cmii-suav-platform-supervisionh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-suav-platform-supervisionh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervisionh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-share + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-share + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-share:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-share + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-share + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-share + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-splice + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-splice + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-splice:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-splice + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-splice + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-splice + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-pangu + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-oms + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-oms + image: harbor.cdcyy.com.cn/cmii/cmii-uav-platform-oms:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: uavcloud-dev + - name: APPLICATION_NAME + value: cmii-uav-platform-oms + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-oms + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-oms + namespace: uavcloud-dev + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 diff --git a/agent-deploy/uavcloud-dev/old/k8s-ingress.yaml b/agent-deploy/uavcloud-dev/old/k8s-ingress.yaml new file mode 100644 index 0000000..21dc54e --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-ingress.yaml @@ -0,0 +1,544 @@ +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: uavcloud-dev + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.uavcloud-dev.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: uavcloud-dev + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-uavcloud-dev.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: uavcloud-dev + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.uavcloud-dev.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 diff --git a/agent-deploy/uavcloud-dev/old/k8s-mongo.yaml b/agent-deploy/uavcloud-dev/old/k8s-mongo.yaml new file mode 100644 index 0000000..a40f2a1 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-mongo.yaml @@ -0,0 +1,75 @@ +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: harbor.cdcyy.com.cn/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- diff --git a/agent-deploy/uavcloud-dev/old/k8s-mysql.yaml b/agent-deploy/uavcloud-dev/old/k8s-mysql.yaml new file mode 100644 index 0000000..1834d25 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-mysql.yaml @@ -0,0 +1,421 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: harbor.cdcyy.com.cn/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv diff --git a/agent-deploy/uavcloud-dev/old/k8s-nacos.yaml b/agent-deploy/uavcloud-dev/old/k8s-nacos.yaml new file mode 100644 index 0000000..6326e2a --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-nacos.yaml @@ -0,0 +1,130 @@ +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: uavcloud-dev + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: harbor.cdcyy.com.cn/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + - containerPort: 9848 + name: tcp-9848 + - containerPort: 9849 + name: tcp-9849 + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- diff --git a/agent-deploy/uavcloud-dev/old/k8s-pvc.yaml b/agent-deploy/uavcloud-dev/old/k8s-pvc.yaml new file mode 100644 index 0000000..cce0138 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-pvc.yaml @@ -0,0 +1,76 @@ +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi diff --git a/agent-deploy/uavcloud-dev/old/k8s-rabbitmq.yaml b/agent-deploy/uavcloud-dev/old/k8s-rabbitmq.yaml new file mode 100644 index 0000000..75ac663 --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-rabbitmq.yaml @@ -0,0 +1,327 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: uavcloud-dev + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: harbor.cdcyy.com.cn/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq diff --git a/agent-deploy/uavcloud-dev/old/k8s-redis.yaml b/agent-deploy/uavcloud-dev/old/k8s-redis.yaml new file mode 100644 index 0000000..5d8461c --- /dev/null +++ b/agent-deploy/uavcloud-dev/old/k8s-redis.yaml @@ -0,0 +1,582 @@ +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: harbor.cdcyy.com.cn/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: uavcloud-dev + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: uavcloud-dev + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: harbor.cdcyy.com.cn/cmii/bitnami-shell:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.uavcloud-dev.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + diff --git a/agent-deploy/z_5.5.0/k8s-backend.yaml b/agent-deploy/z_5.5.0/k8s-backend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-configmap.yaml b/agent-deploy/z_5.5.0/k8s-configmap.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-dashboard.yaml b/agent-deploy/z_5.5.0/k8s-dashboard.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-emqx.yaml b/agent-deploy/z_5.5.0/k8s-emqx.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-frontend.yaml b/agent-deploy/z_5.5.0/k8s-frontend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-ingress.yaml b/agent-deploy/z_5.5.0/k8s-ingress.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-mongo.yaml b/agent-deploy/z_5.5.0/k8s-mongo.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-mysql.yaml b/agent-deploy/z_5.5.0/k8s-mysql.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-nacos.yaml b/agent-deploy/z_5.5.0/k8s-nacos.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-nfs-test.yaml b/agent-deploy/z_5.5.0/k8s-nfs-test.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-nfs.yaml b/agent-deploy/z_5.5.0/k8s-nfs.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-pvc.yaml b/agent-deploy/z_5.5.0/k8s-pvc.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-rabbitmq.yaml b/agent-deploy/z_5.5.0/k8s-rabbitmq.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-redis.yaml b/agent-deploy/z_5.5.0/k8s-redis.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_5.5.0/k8s-srs.yaml b/agent-deploy/z_5.5.0/k8s-srs.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_bjtg/k8s-backend.yaml b/agent-deploy/z_bjtg/k8s-backend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_bjtg/k8s-configmap.yaml b/agent-deploy/z_bjtg/k8s-configmap.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_bjtg/k8s-frontend.yaml b/agent-deploy/z_bjtg/k8s-frontend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_dep/G.go b/agent-deploy/z_dep/G.go old mode 100644 new mode 100755 index 918d63f..4db1446 --- a/agent-deploy/z_dep/G.go +++ b/agent-deploy/z_dep/G.go @@ -4,47 +4,45 @@ import ( "bytes" "fmt" "github.com/go-playground/validator/v10" - "os" - "runtime" "text/template" "wdd.io/agent-common/assert" "wdd.io/agent-common/logger" "wdd.io/agent-common/utils" ) -var ApplyFilePrefix = "" +var ( + K8sDashboardApplyFilePath = "" + EmqxApplyFilePath = "" + MongoApplyFilePath = "" + RabbitMQApplyFilePath = "" + RedisApplyFilePath = "" + MySQLApplyFilePath = "" + NacosApplyFilePath = "" + PVCApplyFilePath = "" + NfsApplyFilePath = "" + NfsTestApplyFilePath = "" + BackendApplyFilePath = "" + FrontendApplyFilePath = "" + SRSApplyFilePath = "" + IngresApplyFilePath = "" + ConfigMapApplyFilePath = "" -var Asserter = assert.Asserter - -var log = logger.Log - -func init() { - switch runtime.GOOS { - case "linux": - ApplyFilePrefix = "/home/wdd/IdeaProjects/ProjectOctopus/agent-deploy/z_bjtg/" - case "darwin": // macOS - ApplyFilePrefix = "MacOS value" - case "windows": - ApplyFilePrefix = "C:\\Users\\wddsh\\Documents\\IdeaProjects\\ProjectOctopus\\agent-deploy\\z_xjyd\\" - default: - ApplyFilePrefix = "Unknown OS" - } - - _ = os.Mkdir(ApplyFilePrefix, os.ModePerm) - fmt.Printf("ApplyFilePrefix: %s\n", ApplyFilePrefix) -} + log = logger.Log + Asserter = assert.Asserter +) type CommonEnvironmentConfig struct { - WebIP string `json:"web_ip,omitempty" valid:"required"` //A1C1IP - WebPort string `json:"web_port,omitempty" valid:"required"` //A1C1JS - HarborIP string `json:"harbor_ip,omitempty" valid:"required"` //A1C2IP - HarborPort string `json:"harbor_port,omitempty" valid:"required"` // default 8033 - Namespace string `json:"namespace,omitempty" valid:"required"` // SUPREME - TagVersion string `json:"tag_version,omitempty" valid:"required"` // KIMMY - TenantEnv string `json:"tenant_env,omitempty"` // TENANT_ENV 只在内部使用 - MinioPublicIP string `json:"minio_public_ip,omitempty"` // M2C1IP - MinioInnerIP string `json:"minio_inner_ip,omitempty"` // M2D2IP - NFSServerIP string `json:"nfs_server_ip,omitempty"` // N1C2IP + WebIP string `json:"web_ip,omitempty" valid:"required"` //A1C1IP + WebPort string `json:"web_port,omitempty" valid:"required"` //A1C1JS + HarborIP string `json:"harbor_ip,omitempty" valid:"required"` //A1C2IP + HarborPort string `json:"harbor_port,omitempty" valid:"required"` // default 8033 + Namespace string `json:"namespace,omitempty" valid:"required"` // SUPREME + TagVersion string `json:"tag_version,omitempty" valid:"required"` // KIMMY + TenantEnv string `json:"tenant_env,omitempty"` // TENANT_ENV 只在内部使用 + MinioPublicIP string `json:"minio_public_ip,omitempty"` // M2C1IP + MinioInnerIP string `json:"minio_inner_ip,omitempty"` // M2D2IP + NFSServerIP string `json:"nfs_server_ip,omitempty"` // N1C2IP + ApplyFilePrefix string } //func (env *CommonEnvironmentConfig) CompactEnv() { @@ -80,6 +78,29 @@ func (env *CommonEnvironmentConfig) ParseCommonEnvToApplyFile(applyTemplate stri return ParseEnvToApplyFile(env, applyTemplate, applyFilePath) } +func (env *CommonEnvironmentConfig) GenerateApplyFilePath() { + + Asserter.NotEmpty(env.ApplyFilePrefix, "apply file prefix is empty!") + + K8sDashboardApplyFilePath = env.ApplyFilePrefix + "k8s-dashboard.yaml" + EmqxApplyFilePath = env.ApplyFilePrefix + "k8s-emqx.yaml" + MongoApplyFilePath = env.ApplyFilePrefix + "k8s-mongo.yaml" + RabbitMQApplyFilePath = env.ApplyFilePrefix + "k8s-rabbitmq.yaml" + RedisApplyFilePath = env.ApplyFilePrefix + "k8s-redis.yaml" + MySQLApplyFilePath = env.ApplyFilePrefix + "k8s-mysql.yaml" + NacosApplyFilePath = env.ApplyFilePrefix + "k8s-nacos.yaml" + PVCApplyFilePath = env.ApplyFilePrefix + "k8s-pvc.yaml" + NfsApplyFilePath = env.ApplyFilePrefix + "k8s-nfs.yaml" + NfsTestApplyFilePath = env.ApplyFilePrefix + "k8s-nfs-test.yaml" + + BackendApplyFilePath = env.ApplyFilePrefix + "k8s-backend.yaml" + FrontendApplyFilePath = env.ApplyFilePrefix + "k8s-frontend.yaml" + SRSApplyFilePath = env.ApplyFilePrefix + "k8s-srs.yaml" + IngresApplyFilePath = env.ApplyFilePrefix + "k8s-ingress.yaml" + ConfigMapApplyFilePath = env.ApplyFilePrefix + "k8s-configmap.yaml" + +} + func ParseEnvToApplyFile(environment any, applyTemplate string, applyFilePath string) bool { randomString := utils.GenerateRandomString(8) diff --git a/agent-deploy/z_file/k8s-backend.yaml b/agent-deploy/z_file/k8s-backend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-configmap.yaml b/agent-deploy/z_file/k8s-configmap.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-dashboard.yaml b/agent-deploy/z_file/k8s-dashboard.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-emqx.yaml b/agent-deploy/z_file/k8s-emqx.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-frontend.yaml b/agent-deploy/z_file/k8s-frontend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-ingress.yaml b/agent-deploy/z_file/k8s-ingress.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-mongo.yaml b/agent-deploy/z_file/k8s-mongo.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-mysql.yaml b/agent-deploy/z_file/k8s-mysql.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-nacos.yaml b/agent-deploy/z_file/k8s-nacos.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-nfs-test.yaml b/agent-deploy/z_file/k8s-nfs-test.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-nfs.yaml b/agent-deploy/z_file/k8s-nfs.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-pvc.yaml b/agent-deploy/z_file/k8s-pvc.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-rabbitmq.yaml b/agent-deploy/z_file/k8s-rabbitmq.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-redis.yaml b/agent-deploy/z_file/k8s-redis.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_file/k8s-srs.yaml b/agent-deploy/z_file/k8s-srs.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-backend.yaml b/agent-deploy/z_xjyd/k8s-backend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-configmap.yaml b/agent-deploy/z_xjyd/k8s-configmap.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-dashboard.yaml b/agent-deploy/z_xjyd/k8s-dashboard.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-emqx.yaml b/agent-deploy/z_xjyd/k8s-emqx.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-frontend.yaml b/agent-deploy/z_xjyd/k8s-frontend.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-ingress.yaml b/agent-deploy/z_xjyd/k8s-ingress.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-mongo.yaml b/agent-deploy/z_xjyd/k8s-mongo.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-mysql.yaml b/agent-deploy/z_xjyd/k8s-mysql.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-nacos.yaml b/agent-deploy/z_xjyd/k8s-nacos.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-nfs-test.yaml b/agent-deploy/z_xjyd/k8s-nfs-test.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-nfs.yaml b/agent-deploy/z_xjyd/k8s-nfs.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-pvc.yaml b/agent-deploy/z_xjyd/k8s-pvc.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-rabbitmq.yaml b/agent-deploy/z_xjyd/k8s-rabbitmq.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-redis.yaml b/agent-deploy/z_xjyd/k8s-redis.yaml old mode 100644 new mode 100755 diff --git a/agent-deploy/z_xjyd/k8s-srs.yaml b/agent-deploy/z_xjyd/k8s-srs.yaml old mode 100644 new mode 100755 diff --git a/agent-go/a_agent/AgentHandler.go b/agent-go/a_agent/AgentHandler.go old mode 100644 new mode 100755 diff --git a/agent-go/a_agent/AgentServerInfo.go b/agent-go/a_agent/AgentServerInfo.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/AppFunction.go b/agent-go/a_executor/AppFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/BaseFunction.go b/agent-go/a_executor/BaseFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/BaseFunction_test.go b/agent-go/a_executor/BaseFunction_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/BasicFunction.go b/agent-go/a_executor/BasicFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/BasicFunction_test.go b/agent-go/a_executor/BasicFunction_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/CommandExecutor.go b/agent-go/a_executor/CommandExecutor.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/FunctionalExecutor.go b/agent-go/a_executor/FunctionalExecutor.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/FunctionalExecutor_test.go b/agent-go/a_executor/FunctionalExecutor_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/HarborExecutor.go b/agent-go/a_executor/HarborExecutor.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/ImageFunction.go b/agent-go/a_executor/ImageFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/ImageFunction_test.go b/agent-go/a_executor/ImageFunction_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/InitFunction.go b/agent-go/a_executor/InitFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/K8sFunction.go b/agent-go/a_executor/K8sFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/MySqlFunction.go b/agent-go/a_executor/MySqlFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/MySqlFunction_test.go b/agent-go/a_executor/MySqlFunction_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/WindowsFunction.go b/agent-go/a_executor/WindowsFunction.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/WindowsFunction_test.go b/agent-go/a_executor/WindowsFunction_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/beans/DockerDaemonConfig.go b/agent-go/a_executor/beans/DockerDaemonConfig.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/beans/HarborTemplateConfig.go b/agent-go/a_executor/beans/HarborTemplateConfig.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/beans/SshKeyConfig.go b/agent-go/a_executor/beans/SshKeyConfig.go old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/script/123 b/agent-go/a_executor/script/123 old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/script/1_node_important.sh b/agent-go/a_executor/script/1_node_important.sh old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/script/install_golang_on_host.sh b/agent-go/a_executor/script/install_golang_on_host.sh old mode 100644 new mode 100755 diff --git a/agent-go/a_executor/script/shutdownFirewall.txt b/agent-go/a_executor/script/shutdownFirewall.txt old mode 100644 new mode 100755 diff --git a/agent-go/a_init/AgentInitialization.go b/agent-go/a_init/AgentInitialization.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/BastionInitializaion.go b/agent-go/a_init/BastionInitializaion.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/ConfigParser.go b/agent-go/a_init/ConfigParser.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/NacosInitalizationDeprecated.go b/agent-go/a_init/NacosInitalizationDeprecated.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/bastion_init/123.txt b/agent-go/a_init/bastion_init/123.txt old mode 100644 new mode 100755 diff --git a/agent-go/a_init/bastion_init/BastionHelp.go b/agent-go/a_init/bastion_init/BastionHelp.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/bastion_init/TrieSearch.go b/agent-go/a_init/bastion_init/TrieSearch.go old mode 100644 new mode 100755 diff --git a/agent-go/a_init/bastion_init/TrieSearch_test.go b/agent-go/a_init/bastion_init/TrieSearch_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/CPU.go b/agent-go/a_status/CPU.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/CPU_test.go b/agent-go/a_status/CPU_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Disk.go b/agent-go/a_status/Disk.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Disk_test.go b/agent-go/a_status/Disk_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Docker_test.go b/agent-go/a_status/Docker_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Dokcer.go b/agent-go/a_status/Dokcer.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Host.go b/agent-go/a_status/Host.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Host_test.go b/agent-go/a_status/Host_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Memory.go b/agent-go/a_status/Memory.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Memory_test.go b/agent-go/a_status/Memory_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Network.go b/agent-go/a_status/Network.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Network_test.go b/agent-go/a_status/Network_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Status.go b/agent-go/a_status/Status.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/Status_test.go b/agent-go/a_status/Status_test.go old mode 100644 new mode 100755 diff --git a/agent-go/a_status/tmp/AgentInfo.json b/agent-go/a_status/tmp/AgentInfo.json old mode 100644 new mode 100755 diff --git a/agent-go/a_status/tmp/AgentStatus.json b/agent-go/a_status/tmp/AgentStatus.json old mode 100644 new mode 100755 diff --git a/agent-go/g/global.go b/agent-go/g/global.go old mode 100644 new mode 100755 diff --git a/agent-go/go.mod b/agent-go/go.mod old mode 100644 new mode 100755 diff --git a/agent-go/go.sum b/agent-go/go.sum old mode 100644 new mode 100755 diff --git a/agent-go/main-dev.go b/agent-go/main-dev.go old mode 100644 new mode 100755 diff --git a/agent-go/main.go b/agent-go/main.go old mode 100644 new mode 100755 diff --git a/agent-go/octopus-agent-cmii.yaml b/agent-go/octopus-agent-cmii.yaml old mode 100644 new mode 100755 diff --git a/agent-go/octopus-agent-local.yaml b/agent-go/octopus-agent-local.yaml old mode 100644 new mode 100755 diff --git a/agent-go/octopus-agent-shanghai.yaml b/agent-go/octopus-agent-shanghai.yaml old mode 100644 new mode 100755 diff --git a/agent-go/octopus-agent-standard.yaml b/agent-go/octopus-agent-standard.yaml old mode 100644 new mode 100755 diff --git a/agent-go/rabbitmq/OctopusMessage.go b/agent-go/rabbitmq/OctopusMessage.go old mode 100644 new mode 100755 diff --git a/agent-go/rabbitmq/OctopusMessageHandler.go b/agent-go/rabbitmq/OctopusMessageHandler.go old mode 100644 new mode 100755 diff --git a/agent-go/rabbitmq/RabbitMsgQueue.go b/agent-go/rabbitmq/RabbitMsgQueue.go old mode 100644 new mode 100755 diff --git a/agent-go/readme.md b/agent-go/readme.md old mode 100644 new mode 100755 diff --git a/agent-go/server-env.yaml b/agent-go/server-env.yaml old mode 100644 new mode 100755 diff --git a/agent-go/shell/agent-bootup.sh b/agent-go/shell/agent-bootup.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/base-agent-sync-offline.sh b/agent-go/shell/base-agent-sync-offline.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/base-minio-sync.sh b/agent-go/shell/base-minio-sync.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/octopus-env.sh b/agent-go/shell/octopus-env.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/octopus-update.sh b/agent-go/shell/octopus-update.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/sync-minio-to-target.sh b/agent-go/shell/sync-minio-to-target.sh old mode 100644 new mode 100755 diff --git a/agent-go/shell/update-app-tag.sh b/agent-go/shell/update-app-tag.sh old mode 100644 new mode 100755 diff --git a/agent-go/tmp/1.sh b/agent-go/tmp/1.sh old mode 100644 new mode 100755 diff --git a/agent-go/tmp/Exsi-copy-virtual-host.sh b/agent-go/tmp/Exsi-copy-virtual-host.sh old mode 100644 new mode 100755 diff --git a/agent-go/tmp/cmii-backend-pod.yaml b/agent-go/tmp/cmii-backend-pod.yaml old mode 100644 new mode 100755 diff --git a/agent-go/tmp/cmii-backend-replicaset.yaml b/agent-go/tmp/cmii-backend-replicaset.yaml old mode 100644 new mode 100755 diff --git a/agent-go/tmp/cmii-frontend-app.yaml b/agent-go/tmp/cmii-frontend-app.yaml old mode 100644 new mode 100755 diff --git a/agent-go/tmp/cmii-frontend-pod.yaml b/agent-go/tmp/cmii-frontend-pod.yaml old mode 100644 new mode 100755 diff --git a/agent-go/tmp/cmii-node-info.yaml b/agent-go/tmp/cmii-node-info.yaml old mode 100644 new mode 100755 diff --git a/agent-go/tmp/executor-om-multiline.json b/agent-go/tmp/executor-om-multiline.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/executor-om-pipeline.json b/agent-go/tmp/executor-om-pipeline.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/executor-om-single.json b/agent-go/tmp/executor-om-single.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/init-from-server-message.json b/agent-go/tmp/init-from-server-message.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/init-octopus-message.json b/agent-go/tmp/init-octopus-message.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/init-to-server-message.json b/agent-go/tmp/init-to-server-message.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/nacos_config_export_20230330143045.zip b/agent-go/tmp/nacos_config_export_20230330143045.zip old mode 100644 new mode 100755 diff --git a/agent-go/tmp/rapid-api.sh b/agent-go/tmp/rapid-api.sh old mode 100644 new mode 100755 diff --git a/agent-go/tmp/status-agentStatus.json b/agent-go/tmp/status-agentStatus.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/status-ping.json b/agent-go/tmp/status-ping.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/test-centos-agent-info.json b/agent-go/tmp/test-centos-agent-info.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/test.go b/agent-go/tmp/test.go old mode 100644 new mode 100755 diff --git a/agent-go/tmp/test.json b/agent-go/tmp/test.json old mode 100644 new mode 100755 diff --git a/agent-go/tmp/test.yaml b/agent-go/tmp/test.yaml old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiDeployOperator.go b/agent-operator/CmiiDeployOperator.go old mode 100644 new mode 100755 index 9cf8945..dad0e53 --- a/agent-operator/CmiiDeployOperator.go +++ b/agent-operator/CmiiDeployOperator.go @@ -1,10 +1,11 @@ package main import ( + "fmt" "os" "path/filepath" "wdd.io/agent-common/utils" - "wdd.io/agent-deploy" + agent_deploy "wdd.io/agent-deploy" "wdd.io/agent-operator/image" ) @@ -36,13 +37,18 @@ func CmiiEnvDeploy() { utils.BeautifulPrintWithTitle(srsMap, "srsMap") // generate and get all old stuff - applyYamlFolder := agent_deploy.CmiiEnvironmentDeploy(shouldDoCompleteDeploy, cmiiNamespace, frontendMap, backendMap) + applyYamlFolder := agent_deploy.CmiiEnvironmentDeploy(shouldDoCompleteDeploy, cmiiNamespace, backendMap, frontendMap) + + fmt.Println(applyYamlFolder) + + // test + //GetNodeWideByKubectl(cmiiNamespace) // clear old apply file - clearOldApplyStuff(applyYamlFolder+"old/", cmiiNamespace) + //clearOldApplyStuff(applyYamlFolder+"old/", cmiiNamespace) // apply new app - applyNewAppStuff(applyYamlFolder, cmiiNamespace) + //applyNewAppStuff(applyYamlFolder, cmiiNamespace) } @@ -71,6 +77,9 @@ func clearOldApplyStuff(oldApplyYamlFolder string, cmiiEnv string) bool { return false } + // rearrange + rearrangeCmiiDeploySequence(files) + for _, file := range files { if filepath.Ext(file.Name()) == ".yaml" || filepath.Ext(file.Name()) == ".yml" { filePath := filepath.Join(oldApplyYamlFolder, file.Name()) @@ -80,3 +89,24 @@ func clearOldApplyStuff(oldApplyYamlFolder string, cmiiEnv string) bool { return true } + +// rearrangeCmiiDeploySequence 重新排序CmiiDeploy顺序 +func rearrangeCmiiDeploySequence(files []os.DirEntry) { + k8s - backend.yaml + k8s - configmap.yaml + k8s - emqx.yaml + k8s - frontend.yaml + k8s - ingress.yaml + k8s - mongo.yaml + k8s - mysql.yaml + k8s - nacos.yaml + k8s - pvc.yaml + k8s - rabbitmq.yaml + k8s - redis.yaml + for _, file := range files { + if filepath.Ext(file.Name()) == ".yaml" || filepath.Ext(file.Name()) == ".yml" { + filePath := filepath.Join(oldApplyYamlFolder, file.Name()) + DeleteByKubectl(filePath, cmiiEnv) + } + } +} diff --git a/agent-operator/CmiiDeployOperator_test.go b/agent-operator/CmiiDeployOperator_test.go new file mode 100755 index 0000000..dbcc590 --- /dev/null +++ b/agent-operator/CmiiDeployOperator_test.go @@ -0,0 +1,8 @@ +package main + +import "testing" + +func TestCmiiEnvDeploy(t *testing.T) { + + CmiiEnvDeploy() +} diff --git a/agent-operator/CmiiK8sOperator.go b/agent-operator/CmiiK8sOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiK8sOperator_test.go b/agent-operator/CmiiK8sOperator_test.go old mode 100644 new mode 100755 index 23d7c43..d2e4a92 --- a/agent-operator/CmiiK8sOperator_test.go +++ b/agent-operator/CmiiK8sOperator_test.go @@ -253,7 +253,7 @@ func TestUpdateCmiiDeploymentImageTag(t *testing.T) { // 计算20:00的时间 now := time.Now() - targetTime := time.Date(now.Year(), now.Month(), now.Day(), 13, 40, 00, 0, now.Location()) + targetTime := time.Date(now.Year(), now.Month(), now.Day(), 17, 23, 00, 0, now.Location()) duration := time.Duration(0) @@ -277,7 +277,7 @@ func TestUpdateCmiiDeploymentImageTag(t *testing.T) { appNameTagMap := map[string]string{ //"cmii-uav-multilink": "5.5.0", //"cmii-uav-process": "5.6.0-060601", - "cmii-uav-industrial-portfolio": "5.6.0-061201", + "cmii-uav-industrial-portfolio": "5.6.0-061302", } for appName, newTag := range appNameTagMap { diff --git a/agent-operator/CmiiMinioOperator.go b/agent-operator/CmiiMinioOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiMinioOperator_test.go b/agent-operator/CmiiMinioOperator_test.go old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiOperator.go b/agent-operator/CmiiOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiOperator_test.go b/agent-operator/CmiiOperator_test.go old mode 100644 new mode 100755 diff --git a/agent-operator/CmiiStatus.go b/agent-operator/CmiiStatus.go old mode 100644 new mode 100755 diff --git a/agent-operator/K8sOperator.go b/agent-operator/K8sOperator.go old mode 100644 new mode 100755 index 41fc15a..088ca71 --- a/agent-operator/K8sOperator.go +++ b/agent-operator/K8sOperator.go @@ -60,8 +60,8 @@ func init() { CmiiDevClusterK8sConfig = "C:\\Users\\wddsh\\Documents\\IdeaProjects\\ProjectOctopus\\agent-operator\\config\\cmii-dev-cluster.yaml" CmiiCoreClusterK8sConfig = "C:\\Users\\wddsh\\Documents\\IdeaProjects\\ProjectOctopus\\agent-operator\\config\\cmii-core-cluster.yaml" case "linux": - CmiiDevClusterK8sConfig = "" - CmiiCoreClusterK8sConfig = "" + CmiiDevClusterK8sConfig = "/home/wdd/IdeaProjects/ProjectOctopus/agent-operator/config/cmii-dev-cluster.yaml" + CmiiCoreClusterK8sConfig = "/home/wdd/IdeaProjects/ProjectOctopus/agent-operator/config/cmii-core-cluster.yaml" } } @@ -94,7 +94,7 @@ func DeleteByKubectl(applyFilePath string, cmiiEnv string) bool { kubeconfig := GetK8sConfigByEnv(cmiiEnv) - cmd := exec.Command("/usr/local/bin/kubectl", "delete", "-f", applyFilePath) + cmd := exec.Command("kubectl", "delete", "-f", applyFilePath) cmd.Env = append(os.Environ(), fmt.Sprintf("KUBECONFIG=%s", kubeconfig)) output, err := cmd.CombinedOutput() if err != nil { @@ -110,7 +110,7 @@ func ApplyByKubectl(applyFilePath string, cmiiEnv string) bool { kubeconfig := GetK8sConfigByEnv(cmiiEnv) - cmd := exec.Command("/usr/local/bin/kubectl", "apply", "-f", applyFilePath) + cmd := exec.Command("kubectl", "apply", "-f", applyFilePath) cmd.Env = append(os.Environ(), fmt.Sprintf("KUBECONFIG=%s", kubeconfig)) output, err := cmd.CombinedOutput() if err != nil { @@ -122,6 +122,20 @@ func ApplyByKubectl(applyFilePath string, cmiiEnv string) bool { return true } +func GetNodeWideByKubectl(cmiiEnv string) { + kubeconfig := GetK8sConfigByEnv(cmiiEnv) + + cmd := exec.Command("kubectl", "get", "nodes", "-o", "wide") + cmd.Env = append(os.Environ(), fmt.Sprintf("KUBECONFIG=%s", kubeconfig)) + output, err := cmd.CombinedOutput() + if err != nil { + log.ErrorF("failed to get nodes: %v\n%s", err, output) + return + } + + fmt.Println(string(output)) +} + func (op *CmiiK8sOperator) checkAndBuildCmiiK8sOperator() { if op.DevClient == nil { diff --git a/agent-operator/K8sOperator_test.go b/agent-operator/K8sOperator_test.go old mode 100644 new mode 100755 diff --git a/agent-operator/config/CmiiImageConfig.go b/agent-operator/config/CmiiImageConfig.go old mode 100644 new mode 100755 diff --git a/agent-operator/config/CmiiK8sConfig.go b/agent-operator/config/CmiiK8sConfig.go old mode 100644 new mode 100755 diff --git a/agent-operator/config/cmii-core-cluster.yaml b/agent-operator/config/cmii-core-cluster.yaml old mode 100644 new mode 100755 diff --git a/agent-operator/config/cmii-dev-cluster.yaml b/agent-operator/config/cmii-dev-cluster.yaml old mode 100644 new mode 100755 diff --git a/agent-operator/config/middleware-amd64.json b/agent-operator/config/middleware-amd64.json old mode 100644 new mode 100755 diff --git a/agent-operator/config/rancher-1.20.4-amd64.json b/agent-operator/config/rancher-1.20.4-amd64.json old mode 100644 new mode 100755 diff --git a/agent-operator/deploy/z_bjtg/k8s-backend.yaml b/agent-operator/deploy/z_bjtg/k8s-backend.yaml new file mode 100755 index 0000000..ce4b99e --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-backend.yaml @@ -0,0 +1,5143 @@ +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-grid-manage + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-manage + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-grid-manage + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-manage + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-grid-manage + image: 10.250.0.110:8033/cmii/cmii-uav-grid-manage:5.1.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-grid-manage + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-grid-manage + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-grid-manage + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-manage + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-grid-manage + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-integration + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-integration + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-integration + image: 10.250.0.110:8033/cmii/cmii-uav-integration:5.5.0-0507 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-integration + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-integration + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-integration + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-integration + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-integration + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-user + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-user + image: 10.250.0.110:8033/cmii/cmii-admin-user:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-admin-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-admin-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-user + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cms + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cms + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cms + image: 10.250.0.110:8033/cmii/cmii-uav-cms:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-cms + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-cms + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cms + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-cms + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cms + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-gateway + image: 10.250.0.110:8033/cmii/cmii-uav-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-logger + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-logger + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-logger + image: 10.250.0.110:8033/cmii/cmii-uav-logger:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-logger + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-logger + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-logger + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-logger + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-logger + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mission + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mission + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mission + image: 10.250.0.110:8033/cmii/cmii-uav-mission:5.5.0-042901 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-mission + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-mission + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mission + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-mission + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mission + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-notice + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-notice + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-notice + image: 10.250.0.110:8033/cmii/cmii-uav-notice:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-notice + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-notice + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-notice + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-notice + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-notice + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-cloud-live + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-cloud-live + image: 10.250.0.110:8033/cmii/cmii-uav-cloud-live:5.5.0-042401 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-cloud-live + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-cloud-live + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-cloud-live + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-cloud-live + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-data-post-process + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-data-post-process + image: 10.250.0.110:8033/cmii/cmii-uav-data-post-process:5.5.0-042501 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-data-post-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-data-post-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-data-post-process + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-data-post-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-grid-datasource + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-datasource + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-grid-datasource + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-datasource + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-grid-datasource + image: 10.250.0.110:8033/cmii/cmii-uav-grid-datasource:5.2.0-24810 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-grid-datasource + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-grid-datasource + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-grid-datasource + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-datasource + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-grid-datasource + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-mqtthandler + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-mqtthandler + image: 10.250.0.110:8033/cmii/cmii-uav-mqtthandler:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-mqtthandler + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-mqtthandler + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-mqtthandler + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-mqtthandler + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-threedsimulation + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-threedsimulation + image: 10.250.0.110:8033/cmii/cmii-uav-threedsimulation:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-threedsimulation + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-threedsimulation + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-threedsimulation + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-threedsimulation + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-tower + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-tower + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-tower + image: 10.250.0.110:8033/cmii/cmii-uav-tower:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-tower + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-tower + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-tower + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-tower + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-tower + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-waypoint + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-waypoint + image: 10.250.0.110:8033/cmii/cmii-uav-waypoint:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-waypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-waypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-waypoint + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-waypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-waypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-supervision + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-suav-supervision + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-supervision + image: 10.250.0.110:8033/cmii/cmii-suav-supervision:5.4.0-032501 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-suav-supervision + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-suav-supervision + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-supervision + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-suav-supervision + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-suav-supervision + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-alarm + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-alarm + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-alarm + image: 10.250.0.110:8033/cmii/cmii-uav-alarm:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-alarm + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-alarm + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-alarm + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-alarm + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-alarm + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-gis-server + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-gis-server + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-gis-server + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-gis-server + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-gis-server + image: 10.250.0.110:8033/cmii/cmii-uav-gis-server:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-gis-server + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-gis-server + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-gis-server + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-gis-server + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-gis-server + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-data + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-data + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-data + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-data + image: 10.250.0.110:8033/cmii/cmii-admin-data:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-admin-data + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-admin-data + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-data + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-data + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-data + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-developer + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-developer + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-developer + image: 10.250.0.110:8033/cmii/cmii-uav-developer:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-developer + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-developer + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-developer + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-developer + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-developer + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-multilink + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-multilink + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-multilink + image: 10.250.0.110:8033/cmii/cmii-uav-multilink:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-multilink + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-multilink + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-multilink + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-multilink + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-multilink + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-material-warehouse + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-material-warehouse + image: 10.250.0.110:8033/cmii/cmii-uav-material-warehouse:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-material-warehouse + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-material-warehouse + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-material-warehouse + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-material-warehouse + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-oauth + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-oauth + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-oauth + image: 10.250.0.110:8033/cmii/cmii-uav-oauth:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-oauth + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-oauth + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-oauth + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-oauth + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-oauth + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-process + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-process + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-process + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-process + image: 10.250.0.110:8033/cmii/cmii-uav-process:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-process + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-process + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-process + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-process + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-process + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-admin-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-admin-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-admin-gateway + image: 10.250.0.110:8033/cmii/cmii-admin-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-admin-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-admin-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-admin-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-admin-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-admin-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-grid-engine + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-engine + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-grid-engine + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-engine + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-grid-engine + image: 10.250.0.110:8033/cmii/cmii-uav-grid-engine:5.1.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-grid-engine + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-grid-engine + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-grid-engine + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-grid-engine + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-grid-engine + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-industrial-portfolio + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-industrial-portfolio + image: 10.250.0.110:8033/cmii/cmii-uav-industrial-portfolio:5.5.0-050801 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-industrial-portfolio + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-industrial-portfolio + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-industrial-portfolio + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-industrial-portfolio + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-device + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-device + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-device + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-device + image: 10.250.0.110:8033/cmii/cmii-uav-device:5.5.0-042301 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-device + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-device + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-device + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-device + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-device + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-app-release + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-app-release + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-app-release + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-app-release + image: 10.250.0.110:8033/cmii/cmii-app-release:4.2.0-validation + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-app-release + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-app-release + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-app-release + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-app-release + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-app-release + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-open-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-open-gateway + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-open-gateway + image: 10.250.0.110:8033/cmii/cmii-open-gateway:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-open-gateway + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-open-gateway + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-open-gateway + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-open-gateway + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-open-gateway + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-brain + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-brain + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-brain + image: 10.250.0.110:8033/cmii/cmii-uav-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-brain + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-brain + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-brain + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-brain + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-brain + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-kpi-monitor + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-kpi-monitor + image: 10.250.0.110:8033/cmii/cmii-uav-kpi-monitor:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-kpi-monitor + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-kpi-monitor + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-kpi-monitor + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-kpi-monitor + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-surveillance + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-surveillance + image: 10.250.0.110:8033/cmii/cmii-uav-surveillance:5.5.0-042901 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-surveillance + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-surveillance + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-surveillance + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-surveillance + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-surveillance + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-user + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-user + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-user + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-user + image: 10.250.0.110:8033/cmii/cmii-uav-user:5.5.0-042801 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-user + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-user + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-user + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-user + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-user + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-airspace + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-airspace + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-airspace + image: 10.250.0.110:8033/cmii/cmii-uav-airspace:5.5.0-042401 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-airspace + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-airspace + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-airspace + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-airspace + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-airspace + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-autowaypoint + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-autowaypoint + image: 10.250.0.110:8033/cmii/cmii-uav-autowaypoint:4.2.0-beta + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-autowaypoint + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-autowaypoint + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-autowaypoint + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-autowaypoint + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-emergency + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: backend + cmii.app: cmii-uav-emergency + template: + metadata: + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + spec: + affinity: + nodeAffinity: + requiredDuringSchedulingIgnoredDuringExecution: + nodeSelectorTerms: + - matchExpressions: + - key: uavcloud.env + operator: In + values: + - demo + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-emergency + image: 10.250.0.110:8033/cmii/cmii-uav-emergency:5.3.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-emergency + - name: CUST_JAVA_OPTS + value: "-Xms200m -Xmx1500m -Dlog4j2.formatMsgNoLookups=true" + - name: NACOS_REGISTRY + value: "helm-nacos:8848" + - name: NACOS_DISCOVERY_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: NACOS_DISCOVERY_PORT + value: "8080" + - name: BIZ_CONFIG_GROUP + value: 5.5.0 + - name: SYS_CONFIG_GROUP + value: 5.5.0 + - name: IMAGE_VERSION + value: 5.5.0 + - name: NACOS_USERNAME + value: "developer" + - name: NACOS_PASSWORD + value: "Deve@9128201" + ports: + - name: pod-port + containerPort: 8080 + protocol: TCP + resources: + limits: + memory: 2Gi + cpu: "2" + requests: + memory: 200Mi + cpu: 200m + livenessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + startupProbe: + httpGet: + path: /cmii/ping + port: pod-port + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 3 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 5 + volumeMounts: + - name: nfs-backend-log-volume + mountPath: /cmii/logs + readOnly: false + subPath: bjtg/cmii-uav-emergency + volumes: + - name: nfs-backend-log-volume + persistentVolumeClaim: + claimName: nfs-backend-log-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-emergency + namespace: bjtg + labels: + cmii.type: backend + cmii.app: cmii-uav-emergency + octopus/control: backend-app-1.0.0 + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/app-version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: backend + cmii.app: cmii-uav-emergency + ports: + - name: backend-tcp + port: 8080 + protocol: TCP + targetPort: 8080 diff --git a/agent-operator/deploy/z_bjtg/k8s-configmap.yaml b/agent-operator/deploy/z_bjtg/k8s-configmap.yaml new file mode 100755 index 0000000..43eba92 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-configmap.yaml @@ -0,0 +1,336 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-cmsportal + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "cmsportal", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-detection + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "detection", + AppClientId: "APP_FDHW2VLVDWPnnOCy" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-emergency + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "emergency", + AppClientId: "APP_aGsTAY1uMZrpKdfk" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-multiterminal + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "multiterminal", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-oms + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "oms", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-traffic + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "traffic", + AppClientId: "APP_Jc8i2wOQ1t73QEJS" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-jiangsuwenlv + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "jiangsuwenlv", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-ai-brain + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "ai-brain", + AppClientId: "APP_rafnuCAmBESIVYMH" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-armypeople + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "armypeople", + AppClientId: "APP_UIegse6Lfou9pO1U" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-media + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "media", + AppClientId: "APP_4AU8lbifESQO4FD6" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-mws + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "mws", + AppClientId: "APP_uKniXPELlRERBBwK" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-open + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "open", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-qinghaitourism + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "qinghaitourism", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-platform + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "platform", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-splice + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "splice", + AppClientId: "APP_zE0M3sTRXrCIJS8Y" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-logistics + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "logistics", + AppClientId: "APP_PvdfRRRBPL8xbIwl" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervisionh5 + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "supervisionh5", + AppClientId: "APP_qqSu82THfexI8PLM" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-base + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "base", + AppClientId: "APP_9LY41OaKSqk2btY0" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-security + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "security", + AppClientId: "APP_JUSEMc7afyWXxvE7" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-securityh5 + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "securityh5", + AppClientId: "APP_N3ImO0Ubfu9peRHD" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-seniclive + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "seniclive", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-share + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "share", + AppClientId: "APP_4lVSVI0ZGxTssir8" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-threedsimulation + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "threedsimulation", + AppClientId: "empty" + } +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: tenant-prefix-supervision + namespace: bjtg +data: + ingress-config.js: |- + var __GlobalIngressConfig = { + TenantEnvironment: "", + CloudHOST: "10.250.0.110:8888", + ApplicationShortName: "supervision", + AppClientId: "APP_qqSu82THfexI8PLM" + } diff --git a/agent-operator/deploy/z_bjtg/k8s-dashboard.yaml b/agent-operator/deploy/z_bjtg/k8s-dashboard.yaml new file mode 100755 index 0000000..430a581 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-dashboard.yaml @@ -0,0 +1,1842 @@ +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system +--- +apiVersion: v1 +kind: Namespace +metadata: + name: kube-system + +--- + +apiVersion: v1 +kind: ServiceAccount +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + kubernetes.io/cluster-service: "true" + name: kubernetes-dashboard + namespace: kube-system +spec: + ports: + - port: 443 + targetPort: 8443 + nodePort: 30554 + selector: + k8s-app: kubernetes-dashboard + type: NodePort + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-certs + namespace: kube-system +type: Opaque + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-csrf + namespace: kube-system +type: Opaque +data: + csrf: "" + +--- + +apiVersion: v1 +kind: Secret +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-key-holder + namespace: kube-system +type: Opaque + +--- + +kind: ConfigMap +apiVersion: v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard-settings + namespace: kube-system + +--- + +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +rules: + # Allow Dashboard to get, update and delete Dashboard exclusive secrets. + - apiGroups: [ "" ] + resources: [ "secrets" ] + resourceNames: [ "kubernetes-dashboard-key-holder", "kubernetes-dashboard-certs", "kubernetes-dashboard-csrf" ] + verbs: [ "get", "update", "delete" ] + # Allow Dashboard to get and update 'kubernetes-dashboard-settings' config map. + - apiGroups: [ "" ] + resources: [ "configmaps" ] + resourceNames: [ "kubernetes-dashboard-settings" ] + verbs: [ "get", "update" ] + # Allow Dashboard to get metrics. + - apiGroups: [ "" ] + resources: [ "services" ] + resourceNames: [ "heapster", "dashboard-metrics-scraper" ] + verbs: [ "proxy" ] + - apiGroups: [ "" ] + resources: [ "services/proxy" ] + resourceNames: [ "heapster", "http:heapster:", "https:heapster:", "dashboard-metrics-scraper", "http:dashboard-metrics-scraper" ] + verbs: [ "get" ] + +--- + +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard +rules: + # Allow Metrics Scraper to get metrics from the Metrics server + - apiGroups: [ "metrics.k8s.io" ] + resources: [ "pods", "nodes" ] + verbs: [ "get", "list", "watch" ] + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: RoleBinding +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: kubernetes-dashboard +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: kubernetes-dashboard +subjects: + - kind: ServiceAccount + name: kubernetes-dashboard + namespace: kube-system + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: kubernetes-dashboard + name: kubernetes-dashboard + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: kubernetes-dashboard + template: + metadata: + labels: + k8s-app: kubernetes-dashboard + spec: + containers: + - name: kubernetes-dashboard + image: 10.250.0.110:8033/cmii/dashboard:v2.0.1 + ports: + - containerPort: 8443 + protocol: TCP + args: + - --auto-generate-certificates + - --namespace=kube-system + # Uncomment the following line to manually specify Kubernetes API server Host + # If not specified, Dashboard will attempt to auto discover the API server and connect + # to it. Uncomment only if the default does not work. + # - --apiserver-host=http://my-address:port + volumeMounts: + - name: kubernetes-dashboard-certs + mountPath: /certs + # Create on-disk volume to store exec logs + - mountPath: /tmp + name: tmp-volume + livenessProbe: + httpGet: + scheme: HTTPS + path: / + port: 8443 + initialDelaySeconds: 30 + timeoutSeconds: 30 + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + volumes: + - name: kubernetes-dashboard-certs + secret: + secretName: kubernetes-dashboard-certs + - name: tmp-volume + emptyDir: { } + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + +--- + +kind: Service +apiVersion: v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + ports: + - port: 8000 + targetPort: 8000 + selector: + k8s-app: dashboard-metrics-scraper + +--- + +kind: Deployment +apiVersion: apps/v1 +metadata: + labels: + k8s-app: dashboard-metrics-scraper + name: dashboard-metrics-scraper + namespace: kube-system +spec: + replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + k8s-app: dashboard-metrics-scraper + template: + metadata: + labels: + k8s-app: dashboard-metrics-scraper + annotations: + seccomp.security.alpha.kubernetes.io/pod: 'runtime/default' + spec: + containers: + - name: dashboard-metrics-scraper + image: 10.250.0.110:8033/cmii/metrics-scraper:v1.0.4 + ports: + - containerPort: 8000 + protocol: TCP + livenessProbe: + httpGet: + scheme: HTTP + path: / + port: 8000 + initialDelaySeconds: 30 + timeoutSeconds: 30 + volumeMounts: + - mountPath: /tmp + name: tmp-volume + securityContext: + allowPrivilegeEscalation: false + readOnlyRootFilesystem: true + runAsUser: 1001 + runAsGroup: 2001 + serviceAccountName: kubernetes-dashboard + # Comment the following tolerations if Dashboard must not be deployed on master + tolerations: + - key: node-role.kubernetes.io/master + effect: NoSchedule + volumes: + - name: tmp-volume + emptyDir: { } +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: admin-user + namespace: kube-system + +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: admin-user +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: cluster-admin +subjects: + - kind: ServiceAccount + name: admin-user + namespace: kube-system diff --git a/agent-operator/deploy/z_bjtg/k8s-emqx.yaml b/agent-operator/deploy/z_bjtg/k8s-emqx.yaml new file mode 100755 index 0000000..2ec8197 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-emqx.yaml @@ -0,0 +1,1578 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-emqxs + namespace: bjtg +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-env + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + EMQX_CLUSTER__K8S__APISERVER: https://kubernetes.default.svc.cluster.local:443 + EMQX_NAME: helm-emqxs + EMQX_CLUSTER__DISCOVERY: k8s + EMQX_CLUSTER__K8S__APP_NAME: helm-emqxs + EMQX_CLUSTER__K8S__SERVICE_NAME: helm-emqxs-headless + EMQX_CLUSTER__K8S__ADDRESS_TYPE: "dns" + EMQX_CLUSTER__K8S__namespace: bjtg + EMQX_CLUSTER__K8S__SUFFIX: svc.cluster.local + EMQX_ALLOW_ANONYMOUS: "false" + EMQX_ACL_NOMATCH: "deny" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-emqxs-cm + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +data: + emqx_auth_username.conf: |- + auth.user.1.username = cmlc + auth.user.1.password = odD8#Ve7.B + auth.user.password_hash = sha256 + + acl.conf: |- + {allow, {user, "admin"}, pubsub, ["admin/#"]}. + {allow, {user, "dashboard"}, subscribe, ["$SYS/#"]}. + {allow, {ipaddr, "127.0.0.1"}, pubsub, ["$SYS/#", "#"]}. + {deny, all, subscribe, ["$SYS/#", {eq, "#"}]}. + {allow, all}. + + loaded_plugins: |- + {emqx_auth_username,true}. + {emqx_management, true}. + {emqx_recon, true}. + {emqx_retainer, false}. + {emqx_dashboard, true}. + {emqx_telemetry, true}. + {emqx_rule_engine, true}. + {emqx_bridge_mqtt, false}. +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + replicas: 1 + serviceName: helm-emqxs-headless + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + template: + metadata: + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + spec: + affinity: { } + serviceAccountName: helm-emqxs + containers: + - name: helm-emqxs + image: 10.250.0.110:8033/cmii/emqx:5.5.1 + imagePullPolicy: Always + ports: + - name: mqtt + containerPort: 1883 + - name: mqttssl + containerPort: 8883 + - name: mgmt + containerPort: 8081 + - name: ws + containerPort: 8083 + - name: wss + containerPort: 8084 + - name: dashboard + containerPort: 18083 + - name: ekka + containerPort: 4370 + envFrom: + - configMapRef: + name: helm-emqxs-env + resources: { } + volumeMounts: + - name: emqx-data + mountPath: "/opt/emqx/data/mnesia" + readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/etc/plugins/emqx_auth_username.conf" + subPath: emqx_auth_username.conf + readOnly: false + # - name: helm-emqxs-cm + # mountPath: "/opt/emqx/etc/acl.conf" + # subPath: "acl.conf" + # readOnly: false + - name: helm-emqxs-cm + mountPath: "/opt/emqx/data/loaded_plugins" + subPath: loaded_plugins + readOnly: false + volumes: + - name: emqx-data + persistentVolumeClaim: + claimName: helm-emqxs + - name: helm-emqxs-cm + configMap: + name: helm-emqxs-cm + items: + - key: emqx_auth_username.conf + path: emqx_auth_username.conf + - key: acl.conf + path: acl.conf + - key: loaded_plugins + path: loaded_plugins +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +rules: + - apiGroups: + - "" + resources: + - endpoints + verbs: + - get + - watch + - list +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-emqxs + namespace: bjtg +subjects: + - kind: ServiceAccount + name: helm-emqxs + namespace: bjtg +roleRef: + kind: Role + name: helm-emqxs + apiGroup: rbac.authorization.k8s.io +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - port: 1883 + name: mqtt + targetPort: 1883 + nodePort: 31883 + - port: 18083 + name: dashboard + targetPort: 18083 + nodePort: 38085 + - port: 8083 + name: mqtt-websocket + targetPort: 8083 + nodePort: 38083 +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-emqxs-headless + namespace: bjtg + labels: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + helm.sh/chart: emqx-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + clusterIP: None + selector: + cmii.type: middleware + cmii.app: helm-emqxs + cmii.emqx.architecture: cluster + ports: + - name: mqtt + port: 1883 + protocol: TCP + targetPort: 1883 + - name: mqttssl + port: 8883 + protocol: TCP + targetPort: 8883 + - name: mgmt + port: 8081 + protocol: TCP + targetPort: 8081 + - name: websocket + port: 8083 + protocol: TCP + targetPort: 8083 + - name: wss + port: 8084 + protocol: TCP + targetPort: 8084 + - name: dashboard + port: 18083 + protocol: TCP + targetPort: 18083 + - name: ekka + port: 4370 + protocol: TCP + targetPort: 4370 diff --git a/agent-operator/deploy/z_bjtg/k8s-frontend.yaml b/agent-operator/deploy/z_bjtg/k8s-frontend.yaml new file mode 100755 index 0000000..960b43d --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-frontend.yaml @@ -0,0 +1,1909 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: nginx-cm + namespace: bjtg + labels: + cmii.type: frontend +data: + nginx.conf: | + user root; + worker_processes auto; + + events { + worker_connections 1024; + } + http { + include mime.types; + default_type application/octet-stream; + + sendfile on; + + keepalive_timeout 600; + + server { + listen 9528; + server_name localhost; + gzip on; + + location / { + root /home/cmii-platform/dist; + index index.html index.htm; + } + + error_page 500 502 503 504 /50x.html; + location = /50x.html { + root html; + } + } + } +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-jiangsuwenlv + image: 10.250.0.110:8033/cmii/cmii-uav-platform-jiangsuwenlv:4.1.3-jiangsu-0427 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-jiangsuwenlv + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-jiangsuwenlv + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-jiangsuwenlv + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-jiangsuwenlv + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-logistics + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-logistics + image: 10.250.0.110:8033/cmii/cmii-uav-platform-logistics:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-logistics + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-logistics + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-logistics + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-logistics + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-multiterminal + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-multiterminal + image: 10.250.0.110:8033/cmii/cmii-uav-platform-multiterminal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-multiterminal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-multiterminal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-multiterminal + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-multiterminal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-open + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-open + image: 10.250.0.110:8033/cmii/cmii-uav-platform-open:5.5.0-0419 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-open + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-open + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-open + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-open + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform + image: 10.250.0.110:8033/cmii/cmii-uav-platform:5.5.0-27712-0507 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-platform + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-ai-brain + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-ai-brain + image: 10.250.0.110:8033/cmii/cmii-uav-platform-ai-brain:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-ai-brain + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-ai-brain + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-ai-brain + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-ai-brain + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-mws + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-mws + image: 10.250.0.110:8033/cmii/cmii-uav-platform-mws:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-mws + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-mws + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-mws + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-mws + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-security + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-security + image: 10.250.0.110:8033/cmii/cmii-uav-platform-security:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-security + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-security + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-security + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-security + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-splice + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-splice + image: 10.250.0.110:8033/cmii/cmii-uav-platform-splice:5.5.0-042601 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-splice + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-splice + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-splice + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-splice + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-threedsimulation + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-threedsimulation + image: 10.250.0.110:8033/cmii/cmii-uav-platform-threedsimulation:5.2.0-21392 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-threedsimulation + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-threedsimulation + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-threedsimulation + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-threedsimulation + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-securityh5 + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-securityh5 + image: 10.250.0.110:8033/cmii/cmii-uav-platform-securityh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-securityh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-securityh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-securityh5 + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-securityh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-share + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-share + image: 10.250.0.110:8033/cmii/cmii-uav-platform-share:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-share + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-share + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-share + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-share + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-armypeople + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-armypeople + image: 10.250.0.110:8033/cmii/cmii-uav-platform-armypeople:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-armypeople + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-armypeople + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-armypeople + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-armypeople + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-cms-portal + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-cms-portal + image: 10.250.0.110:8033/cmii/cmii-uav-platform-cms-portal:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-cms-portal + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-cmsportal + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-cms-portal + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-cms-portal + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-media + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-media + image: 10.250.0.110:8033/cmii/cmii-uav-platform-media:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-media + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-media + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-media + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-media + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-oms + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-oms + image: 10.250.0.110:8033/cmii/cmii-uav-platform-oms:5.5.0-042801 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-oms + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-oms + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-oms + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-oms + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-qinghaitourism + image: 10.250.0.110:8033/cmii/cmii-uav-platform-qinghaitourism:4.1.0-21377-0508 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-qinghaitourism + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-qinghaitourism + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-qinghaitourism + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-qinghaitourism + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervision + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervision + image: 10.250.0.110:8033/cmii/cmii-suav-platform-supervision:5.5.0-042301 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-suav-platform-supervision + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervision + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervision + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervision + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-suav-platform-supervisionh5 + image: 10.250.0.110:8033/cmii/cmii-suav-platform-supervisionh5:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-suav-platform-supervisionh5 + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-supervisionh5 + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-suav-platform-supervisionh5 + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-suav-platform-supervisionh5 + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-base + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-base + image: 10.250.0.110:8033/cmii/cmii-uav-platform-base:5.4.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-base + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-base + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-base + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-base + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: cmii-uav-platform-detection + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/app-version: 5.5.0 +spec: + replicas: 1 + strategy: + rollingUpdate: + maxUnavailable: 1 + selector: + matchLabels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + template: + metadata: + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + spec: + imagePullSecrets: + - name: harborsecret + containers: + - name: cmii-uav-platform-detection + image: 10.250.0.110:8033/cmii/cmii-uav-platform-detection:5.5.0 + imagePullPolicy: Always + env: + - name: K8S_NAMESPACE + value: bjtg + - name: APPLICATION_NAME + value: cmii-uav-platform-detection + ports: + - name: platform-9528 + containerPort: 9528 + protocol: TCP + resources: + limits: + cpu: "1" + memory: 1Gi + requests: + cpu: 500m + memory: 500Mi + volumeMounts: + - name: nginx-conf + mountPath: /usr/local/nginx/conf/nginx.conf + subPath: nginx.conf + - name: tenant-prefix + subPath: ingress-config.js + mountPath: /home/cmii-platform/dist/ingress-config.js + volumes: + - name: nginx-conf + configMap: + name: nginx-cm + items: + - key: nginx.conf + path: nginx.conf + - name: tenant-prefix + configMap: + name: tenant-prefix-detection + items: + - key: ingress-config.js + path: ingress-config.js +--- +apiVersion: v1 +kind: Service +metadata: + name: cmii-uav-platform-detection + namespace: bjtg + labels: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + octopus.control: frontend-app-wdd + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.type: frontend + cmii.app: cmii-uav-platform-detection + ports: + - name: web-svc-port + port: 9528 + protocol: TCP + targetPort: 9528 diff --git a/agent-operator/deploy/z_bjtg/k8s-ingress.yaml b/agent-operator/deploy/z_bjtg/k8s-ingress.yaml new file mode 100755 index 0000000..d58c556 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-ingress.yaml @@ -0,0 +1,3264 @@ +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: frontend-applications-ingress + namespace: bjtg + labels: + type: frontend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + rewrite ^(/green)$ $1/ redirect; + rewrite ^(/supervision)$ $1/ redirect; + rewrite ^(/inspection)$ $1/ redirect; + rewrite ^(/park)$ $1/ redirect; + rewrite ^(/pangu)$ $1/ redirect; + rewrite ^(/ai-brain)$ $1/ redirect; + rewrite ^(/base)$ $1/ redirect; + rewrite ^(/cms)$ $1/ redirect; + rewrite ^(/cmsportal)$ $1/ redirect; + rewrite ^(/detection)$ $1/ redirect; + rewrite ^(/emergency)$ $1/ redirect; + rewrite ^(/hyper)$ $1/ redirect; + rewrite ^(/logistics)$ $1/ redirect; + rewrite ^(/mws)$ $1/ redirect; + rewrite ^(/mws-admin)$ $1/ redirect; + rewrite ^(/oms)$ $1/ redirect; + rewrite ^(/open)$ $1/ redirect; + rewrite ^(/security)$ $1/ redirect; + rewrite ^(/share)$ $1/ redirect; + rewrite ^(/splice)$ $1/ redirect; + rewrite ^(/splice-visual)$ $1/ redirect; + rewrite ^(/traffic)$ $1/ redirect; + rewrite ^(/visualization)$ $1/ redirect; + rewrite ^(/communication)$ $1/ redirect; + rewrite ^(/infrastructure)$ $1/ redirect; + rewrite ^(/media)$ $1/ redirect; + rewrite ^(/seniclive)$ $1/ redirect; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /inspection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /supervision/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervision + servicePort: 9528 + - path: /supervisionh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-platform-supervisionh5 + servicePort: 9528 + - path: /green/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /park/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /pangu/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /emersupport/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /infrastructure/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /ai-brain/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-ai-brain + servicePort: 9528 + - path: /base/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-base + servicePort: 9528 + - path: /cms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms + servicePort: 9528 + - path: /cmsportal/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-cms-portal + servicePort: 9528 + - path: /detection/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-detection + servicePort: 9528 + - path: /emergency/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-emergency-rescue + servicePort: 9528 + - path: /hyper/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-hyperspectral + servicePort: 9528 + - path: /logistics/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-logistics + servicePort: 9528 + - path: /mws/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws + servicePort: 9528 + - path: /mws-admin/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-mws-admin + servicePort: 9528 + - path: /oms/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-oms + servicePort: 9528 + - path: /open/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-open + servicePort: 9528 + - path: /security/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /share/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-share + servicePort: 9528 + - path: /splice/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice + servicePort: 9528 + - path: /splice-visual/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-splice-visual + servicePort: 9528 + - path: /traffic/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /visualization/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-visualization + servicePort: 9528 + - path: /communication/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 + - path: /media/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-media + servicePort: 9528 + - path: /seniclive/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-seniclive + servicePort: 9528 + - path: /jiangsuwenlv/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-jiangsuwenlv + servicePort: 9528 + - path: /qinghaitourism/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-qinghaitourism + servicePort: 9528 + - path: /securityh5/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform-securityh5 + servicePort: 9528 + - path: /fireRescue/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-platform + servicePort: 9528 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: backend-applications-ingress + namespace: bjtg + labels: + type: backend + octopus.control: all-ingress-config-wdd + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" +spec: + rules: + - host: cmii-admin-data.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-data + servicePort: 8080 + - host: cmii-admin-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - host: cmii-admin-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-user + servicePort: 8080 + - host: cmii-open-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - host: cmii-uav-airspace.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-airspace + servicePort: 8080 + - host: cmii-uav-brain.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-brain + servicePort: 8080 + - host: cmii-uav-clusters.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-clusters + servicePort: 8080 + - host: cmii-uav-cms.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cms + servicePort: 8080 + - host: cmii-uav-data-post-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-data-post-process + servicePort: 8080 + - host: cmii-uav-developer.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-developer + servicePort: 8080 + - host: cmii-uav-device.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-device + servicePort: 8080 + - host: cmii-uav-gateway.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 + - host: cmii-uav-industrial-portfolio.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-industrial-portfolio + servicePort: 8080 + - host: cmii-uav-kpi-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-kpi-monitor + servicePort: 8080 + - host: cmii-uav-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-live + servicePort: 8080 + - host: cmii-uav-cloud-live.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-cloud-live + servicePort: 8080 + - host: cmii-uav-logger.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-logger + servicePort: 8080 + - host: cmii-uav-material-warehouse.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-material-warehouse + servicePort: 8080 + - host: cmii-uav-mission.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mission + servicePort: 8080 + - host: cmii-uav-monitor.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-monitor + servicePort: 8080 + - host: cmii-uav-mqtthandler.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-mqtthandler + servicePort: 8080 + - host: cmii-uav-notice.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-notice + servicePort: 8080 + - host: cmii-uav-oauth.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-oauth + servicePort: 8080 + - host: cmii-uav-process.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-process + servicePort: 8080 + - host: cmii-uav-security-system.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-security-system + servicePort: 8080 + - host: cmii-uav-surveillance.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-surveillance + servicePort: 8080 + - host: cmii-uav-user.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-user + servicePort: 8080 + - host: cmii-uav-waypoint.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-waypoint + servicePort: 8080 + - host: cmii-uav-alarm.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-alarm + servicePort: 8080 + - host: cmii-uav-emergency.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-emergency + servicePort: 8080 + - host: cmii-uav-integration.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-integration + servicePort: 8080 + - host: cmii-suav-supervision.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-suav-supervision + servicePort: 8080 + - host: cmii-uav-gis-server.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gis-server + servicePort: 8080 + - host: cmii-uav-grid-datasource.uavcloud-bjtg.io + http: + paths: + - path: / + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-grid-datasource + servicePort: 8080 +--- +apiVersion: networking.k8s.io/v1beta1 +kind: Ingress +metadata: + name: all-gateways-ingress + namespace: bjtg + labels: + type: api-gateway + octopus.control: all-ingress-config-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + kubernetes.io/ingress.class: "nginx" + nginx.ingress.kubernetes.io/enable-cors: "true" + nginx.ingress.kubernetes.io/rewrite-target: /$1 + nginx.ingress.kubernetes.io/configuration-snippet: | + proxy_set_header Upgrade $http_upgrade; + proxy_set_header Connection "Upgrade"; +spec: + rules: + - host: fake-domain.bjtg.io + http: + paths: + - path: /oms/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-admin-gateway + servicePort: 8080 + - path: /open/api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-open-gateway + servicePort: 8080 + - path: /api/?(.*) + pathType: ImplementationSpecific + backend: + serviceName: cmii-uav-gateway + servicePort: 8080 diff --git a/agent-operator/deploy/z_bjtg/k8s-mongo.yaml b/agent-operator/deploy/z_bjtg/k8s-mongo.yaml new file mode 100755 index 0000000..70fd990 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-mongo.yaml @@ -0,0 +1,450 @@ +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + type: ClusterIP + selector: + cmii.app: helm-mongo + cmii.type: middleware + ports: + - port: 27017 + name: server-27017 + targetPort: 27017 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-mongo + replicas: 1 + selector: + matchLabels: + cmii.app: helm-mongo + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-mongo + cmii.type: middleware + helm.sh/chart: mongo-1.1.0 + app.kubernetes.io/managed-by: octopus-control + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: helm-mongo + image: 10.250.0.110:8033/cmii/mongo:5.0 + resources: { } + ports: + - containerPort: 27017 + name: mongo27017 + protocol: TCP + env: + - name: MONGO_INITDB_ROOT_USERNAME + value: cmlc + - name: MONGO_INITDB_ROOT_PASSWORD + value: REdPza8#oVlt + volumeMounts: + - name: mongo-data + mountPath: /data/db + readOnly: false + subPath: default/helm-mongo/data/db + volumes: + - name: mongo-data + persistentVolumeClaim: + claimName: helm-mongo +--- diff --git a/agent-operator/deploy/z_bjtg/k8s-mysql.yaml b/agent-operator/deploy/z_bjtg/k8s-mysql.yaml new file mode 100755 index 0000000..8ed31d1 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-mysql.yaml @@ -0,0 +1,2526 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + annotations: +secrets: + - name: helm-mysql +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + mysql-root-password: "UXpmWFFoZDNiUQ==" + mysql-password: "S0F0cm5PckFKNw==" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + my.cnf: |- + + [mysqld] + port=3306 + basedir=/opt/bitnami/mysql + datadir=/bitnami/mysql/data + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid + socket=/opt/bitnami/mysql/tmp/mysql.sock + log-error=/bitnami/mysql/data/error.log + general_log_file = /bitnami/mysql/data/general.log + slow_query_log_file = /bitnami/mysql/data/slow.log + innodb_data_file_path = ibdata1:512M:autoextend + innodb_buffer_pool_size = 512M + innodb_buffer_pool_instances = 2 + innodb_log_file_size = 512M + innodb_log_files_in_group = 4 + innodb_log_files_in_group = 4 + log-bin = /bitnami/mysql/data/mysql-bin + max_binlog_size=1G + transaction_isolation = REPEATABLE-READ + default_storage_engine = innodb + character-set-server = utf8mb4 + collation-server=utf8mb4_bin + binlog_format = ROW + binlog_rows_query_log_events=on + binlog_cache_size=4M + binlog_expire_logs_seconds = 1296000 + max_binlog_cache_size=2G + gtid_mode = on + enforce_gtid_consistency = 1 + sync_binlog = 1 + innodb_flush_log_at_trx_commit = 1 + innodb_flush_method = O_DIRECT + log_slave_updates=1 + relay_log_recovery = 1 + relay-log-purge = 1 + default_time_zone = '+08:00' + lower_case_table_names=1 + log_bin_trust_function_creators=1 + group_concat_max_len=67108864 + innodb_io_capacity = 4000 + innodb_io_capacity_max = 8000 + innodb_flush_sync = 0 + innodb_flush_neighbors = 0 + innodb_write_io_threads = 8 + innodb_read_io_threads = 8 + innodb_purge_threads = 4 + innodb_page_cleaners = 4 + innodb_open_files = 65535 + innodb_max_dirty_pages_pct = 50 + innodb_lru_scan_depth = 4000 + innodb_checksum_algorithm = crc32 + innodb_lock_wait_timeout = 10 + innodb_rollback_on_timeout = 1 + innodb_print_all_deadlocks = 1 + innodb_file_per_table = 1 + innodb_online_alter_log_max_size = 4G + innodb_stats_on_metadata = 0 + innodb_thread_concurrency = 0 + innodb_sync_spin_loops = 100 + innodb_spin_wait_delay = 30 + lock_wait_timeout = 3600 + slow_query_log = 1 + long_query_time = 10 + log_queries_not_using_indexes =1 + log_throttle_queries_not_using_indexes = 60 + min_examined_row_limit = 100 + log_slow_admin_statements = 1 + log_slow_slave_statements = 1 + default_authentication_plugin=mysql_native_password + skip-name-resolve=1 + explicit_defaults_for_timestamp=1 + plugin_dir=/opt/bitnami/mysql/plugin + max_allowed_packet=128M + max_connections = 2000 + max_connect_errors = 1000000 + table_definition_cache=2000 + table_open_cache_instances=64 + tablespace_definition_cache=1024 + thread_cache_size=256 + interactive_timeout = 600 + wait_timeout = 600 + tmpdir=/opt/bitnami/mysql/tmp + max_allowed_packet=32M + bind-address=0.0.0.0 + performance_schema = 1 + performance_schema_instrument = '%memory%=on' + performance_schema_instrument = '%lock%=on' + innodb_monitor_enable=ALL + + [mysql] + no-auto-rehash + + [mysqldump] + quick + max_allowed_packet = 32M + + [client] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + default-character-set=UTF8 + plugin_dir=/opt/bitnami/mysql/plugin + + [manager] + port=3306 + socket=/opt/bitnami/mysql/tmp/mysql.sock + pid-file=/opt/bitnami/mysql/tmp/mysqld.pid +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-mysql-init-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: primary +data: + create_users_grants_core.sql: |- + create + user zyly@'%' identified by 'Cmii@451315'; + grant select on *.* to zyly@'%'; + create + user zyly_qc@'%' identified by 'Uh)E_owCyb16'; + grant all + on *.* to zyly_qc@'%'; + create + user k8s_admin@'%' identified by 'fP#UaH6qQ3)8'; + grant all + on *.* to k8s_admin@'%'; + create + user audit_dba@'%' identified by 'PjCzqiBmJaTpgkoYXynH'; + grant all + on *.* to audit_dba@'%'; + create + user db_backup@'%' identified by 'RU5Pu(4FGdT9'; + GRANT + SELECT, RELOAD, PROCESS, LOCK TABLES, REPLICATION CLIENT, EVENT + on *.* to db_backup@'%'; + create + user monitor@'%' identified by 'PL3#nGtrWbf-'; + grant REPLICATION + CLIENT on *.* to monitor@'%'; + flush + privileges; +--- +kind: Service +apiVersion: v1 +metadata: + name: cmii-mysql + namespace: bjtg + labels: + app.kubernetes.io/component: primary + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + octopus.control: mysql-db-wdd +spec: + ports: + - name: mysql + protocol: TCP + port: 13306 + targetPort: mysql + selector: + app.kubernetes.io/component: primary + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.app: mysql + cmii.type: middleware + type: ClusterIP +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql-headless + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: ClusterIP + clusterIP: None + publishNotReadyAddresses: true + ports: + - name: mysql + port: 3306 + targetPort: mysql + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + annotations: +spec: + type: NodePort + ports: + - name: mysql + port: 3306 + protocol: TCP + targetPort: mysql + nodePort: 33306 + selector: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-mysql + namespace: bjtg + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: mysql-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + serviceName: helm-mysql + updateStrategy: + type: RollingUpdate + template: + metadata: + annotations: + checksum/configuration: 6b60fa0f3a846a6ada8effdc4f823cf8003d42a8c8f630fe8b1b66d3454082dd + labels: + app.kubernetes.io/name: mysql-db + octopus.control: mysql-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: mysql + app.kubernetes.io/component: primary + spec: + serviceAccountName: helm-mysql + affinity: { } + nodeSelector: + mysql-deploy: "true" + securityContext: + fsGroup: 1001 + initContainers: + - name: change-volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:11-debian-11-r136 + imagePullPolicy: "Always" + command: + - /bin/bash + - -ec + - | + chown -R 1001:1001 /bitnami/mysql + securityContext: + runAsUser: 0 + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + containers: + - name: mysql + image: 10.250.0.110:8033/cmii/mysql:8.1.0-debian-11-r42 + imagePullPolicy: "IfNotPresent" + securityContext: + runAsUser: 1001 + env: + - name: BITNAMI_DEBUG + value: "true" + - name: MYSQL_ROOT_PASSWORD + valueFrom: + secretKeyRef: + name: helm-mysql + key: mysql-root-password + - name: MYSQL_DATABASE + value: "cmii" + ports: + - name: mysql + containerPort: 3306 + livenessProbe: + failureThreshold: 5 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + readinessProbe: + failureThreshold: 5 + initialDelaySeconds: 30 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 3 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + startupProbe: + failureThreshold: 60 + initialDelaySeconds: 120 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + exec: + command: + - /bin/bash + - -ec + - | + password_aux="${MYSQL_ROOT_PASSWORD:-}" + if [[ -f "${MYSQL_ROOT_PASSWORD_FILE:-}" ]]; then + password_aux=$(cat "$MYSQL_ROOT_PASSWORD_FILE") + fi + mysqladmin status -uroot -p"${password_aux}" + resources: + limits: { } + requests: { } + volumeMounts: + - name: mysql-data + mountPath: /bitnami/mysql + - name: custom-init-scripts + mountPath: /docker-entrypoint-initdb.d + - name: config + mountPath: /opt/bitnami/mysql/conf/my.cnf + subPath: my.cnf + volumes: + - name: config + configMap: + name: helm-mysql + - name: custom-init-scripts + configMap: + name: helm-mysql-init-scripts + - name: mysql-data + hostPath: + path: /var/lib/docker/mysql-pv diff --git a/agent-operator/deploy/z_bjtg/k8s-nacos.yaml b/agent-operator/deploy/z_bjtg/k8s-nacos.yaml new file mode 100755 index 0000000..a415b09 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-nacos.yaml @@ -0,0 +1,756 @@ +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-nacos-cm + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +data: + mysql.db.name: "cmii_nacos_config" + mysql.db.host: "helm-mysql" + mysql.port: "3306" + mysql.user: "k8s_admin" + mysql.password: "fP#UaH6qQ3)8" +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + type: NodePort + selector: + cmii.app: helm-nacos + cmii.type: middleware + ports: + - port: 8848 + name: server + targetPort: 8848 + nodePort: 38989 + - port: 9848 + name: server12 + targetPort: 9848 + nodePort: 38912 + - port: 9849 + name: server23 + targetPort: 9849 + nodePort: 38923 +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-nacos + namespace: bjtg + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: Helm + app.kubernetes.io/version: 5.5.0 +spec: + serviceName: helm-nacos + replicas: 1 + selector: + matchLabels: + cmii.app: helm-nacos + cmii.type: middleware + template: + metadata: + labels: + cmii.app: helm-nacos + cmii.type: middleware + octopus.control: nacos-wdd + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/version: 5.5.0 + annotations: + pod.alpha.kubernetes.io/initialized: "true" + spec: + affinity: { } + containers: + - name: nacos-server + image: 10.250.0.110:8033/cmii/nacos-server:v2.1.2 + ports: + - containerPort: 8848 + name: dashboard + env: + - name: NACOS_AUTH_ENABLE + value: "false" + - name: NACOS_REPLICAS + value: "1" + - name: MYSQL_SERVICE_DB_NAME + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.name + - name: MYSQL_SERVICE_PORT + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.port + - name: MYSQL_SERVICE_USER + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.user + - name: MYSQL_SERVICE_PASSWORD + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.password + - name: MYSQL_SERVICE_HOST + valueFrom: + configMapKeyRef: + name: helm-nacos-cm + key: mysql.db.host + - name: NACOS_SERVER_PORT + value: "8848" + - name: NACOS_APPLICATION_PORT + value: "8848" + - name: PREFER_HOST_MODE + value: "hostname" + - name: MODE + value: standalone + - name: SPRING_DATASOURCE_PLATFORM + value: mysql +--- diff --git a/agent-operator/deploy/z_bjtg/k8s-nfs-test.yaml b/agent-operator/deploy/z_bjtg/k8s-nfs-test.yaml new file mode 100755 index 0000000..5ff83c4 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-nfs-test.yaml @@ -0,0 +1,216 @@ +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 +--- +kind: PersistentVolumeClaim +apiVersion: v1 +metadata: + name: test-claim + annotations: + volume.beta.kubernetes.io/storage-class: "nfs-prod-distribute" #与nfs-StorageClass.yaml metadata.name保持一致 +spec: + accessModes: + - ReadWriteOnce + storageClassName: nfs-prod-distribute + resources: + requests: + storage: 1Mi +--- +kind: Pod +apiVersion: v1 +metadata: + name: test-pod +spec: + containers: + - name: test-pod + image: 10.250.0.110:8033/cmii/busybox + command: + - "/bin/sh" + args: + - "-c" + - "touch /mnt/NFS-CREATE-SUCCESS && exit 0 || exit 1" #创建一个SUCCESS文件后退出 + volumeMounts: + - name: nfs-pvc + mountPath: "/mnt" + restartPolicy: "Never" + volumes: + - name: nfs-pvc + persistentVolumeClaim: + claimName: test-claim #与PVC名称保持一致 diff --git a/agent-operator/deploy/z_bjtg/k8s-nfs.yaml b/agent-operator/deploy/z_bjtg/k8s-nfs.yaml new file mode 100755 index 0000000..2a8f10e --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-nfs.yaml @@ -0,0 +1,672 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #根据实际环境设定namespace,下面类同 +--- +kind: ClusterRole +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: nfs-client-provisioner-runner +rules: + - apiGroups: [ "" ] + resources: [ "persistentvolumes" ] + verbs: [ "get", "list", "watch", "create", "delete" ] + - apiGroups: [ "" ] + resources: [ "persistentvolumeclaims" ] + verbs: [ "get", "list", "watch", "update" ] + - apiGroups: [ "storage.k8s.io" ] + resources: [ "storageclasses" ] + verbs: [ "get", "list", "watch" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create", "update", "patch" ] +--- +kind: ClusterRoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: run-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: ClusterRole + # name: nfs-client-provisioner-runner + name: cluster-admin + apiGroup: rbac.authorization.k8s.io +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get", "list", "watch", "create", "update", "patch" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: leader-locking-nfs-client-provisioner +subjects: + - kind: ServiceAccount + name: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system +roleRef: + kind: Role + name: leader-locking-nfs-client-provisioner + apiGroup: rbac.authorization.k8s.io + +--- +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: nfs-prod-distribute +provisioner: cmlc-nfs-storage #这里的名称要和provisioner配置文件中的环境变量PROVISIONER_NAME保持一致parameters: archiveOnDelete: "false" +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: nfs-client-provisioner + labels: + app: nfs-client-provisioner + # replace with namespace where provisioner is deployed + namespace: kube-system #与RBAC文件中的namespace保持一致 +spec: + replicas: 1 + selector: + matchLabels: + app: nfs-client-provisioner + strategy: + type: Recreate + template: + metadata: + labels: + app: nfs-client-provisioner + spec: + serviceAccountName: nfs-client-provisioner + containers: + - name: nfs-client-provisioner + image: 10.250.0.110:8033/cmii/nfs-subdir-external-provisioner:v4.0.2 + volumeMounts: + - name: nfs-client-root + mountPath: /persistentvolumes + env: + - name: PROVISIONER_NAME + value: cmlc-nfs-storage + - name: NFS_SERVER + value: 10.250.0.110 + - name: NFS_PATH + value: /var/lib/docker/nfs_data + volumes: + - name: nfs-client-root + nfs: + server: 10.250.0.110 + path: /var/lib/docker/nfs_data diff --git a/agent-operator/deploy/z_bjtg/k8s-pvc.yaml b/agent-operator/deploy/z_bjtg/k8s-pvc.yaml new file mode 100755 index 0000000..916b9eb --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-pvc.yaml @@ -0,0 +1,456 @@ +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: nfs-backend-log-pvc + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: nfs-backend-log-pvc + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 100Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-emqxs + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-emqxs + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-mongo + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-mongo + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + cmii.type: middleware-base + cmii.app: helm-rabbitmq + helm.sh/chart: all-persistence-volume-claims-1.1.0 + app.kubernetes.io/version: 5.5.0 +spec: + storageClassName: nfs-prod-distribute + accessModes: + - ReadWriteMany + volumeMode: Filesystem + resources: + requests: + storage: 20Gi diff --git a/agent-operator/deploy/z_bjtg/k8s-rabbitmq.yaml b/agent-operator/deploy/z_bjtg/k8s-rabbitmq.yaml new file mode 100755 index 0000000..975e4b5 --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-rabbitmq.yaml @@ -0,0 +1,3924 @@ +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq +--- +apiVersion: v1 +kind: ServiceAccount +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +automountServiceAccountToken: true +secrets: + - name: helm-rabbitmq +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +type: Opaque +data: + rabbitmq-password: "blljUk45MXIuX2hq" + rabbitmq-erlang-cookie: "emFBRmt1ZU1xMkJieXZvdHRYbWpoWk52UThuVXFzcTU=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-rabbitmq-config + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +data: + rabbitmq.conf: |- + ## Username and password + ## + default_user = admin + default_pass = nYcRN91r._hj + ## Clustering + ## + cluster_formation.peer_discovery_backend = rabbit_peer_discovery_k8s + cluster_formation.k8s.host = kubernetes.default.svc.cluster.local + cluster_formation.node_cleanup.interval = 10 + cluster_formation.node_cleanup.only_log_warning = true + cluster_partition_handling = autoheal + # queue master locator + queue_master_locator = min-masters + # enable guest user + loopback_users.guest = false + #default_vhost = default-vhost + #disk_free_limit.absolute = 50MB + #load_definitions = /app/load_definition.json +--- +kind: Role +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +rules: + - apiGroups: [ "" ] + resources: [ "endpoints" ] + verbs: [ "get" ] + - apiGroups: [ "" ] + resources: [ "events" ] + verbs: [ "create" ] +--- +kind: RoleBinding +apiVersion: rbac.authorization.k8s.io/v1 +metadata: + name: helm-rabbitmq-endpoint-reader + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +subjects: + - kind: ServiceAccount + name: helm-rabbitmq +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: Role + name: helm-rabbitmq-endpoint-reader +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq-headless + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + clusterIP: None + ports: + - name: epmd + port: 4369 + targetPort: epmd + - name: amqp + port: 5672 + targetPort: amqp + - name: dist + port: 25672 + targetPort: dist + - name: dashboard + port: 15672 + targetPort: stats + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + publishNotReadyAddresses: true +--- +apiVersion: v1 +kind: Service +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + type: NodePort + ports: + - name: amqp + port: 5672 + targetPort: amqp + nodePort: 35672 + - name: dashboard + port: 15672 + targetPort: dashboard + nodePort: 35675 + selector: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg +--- +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-rabbitmq + namespace: bjtg + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq +spec: + serviceName: helm-rabbitmq-headless + podManagementPolicy: OrderedReady + replicas: 1 + updateStrategy: + type: RollingUpdate + selector: + matchLabels: + app.kubernetes.io/name: helm-rabbitmq + app.kubernetes.io/release: bjtg + template: + metadata: + labels: + app.kubernetes.io/name: helm-rabbitmq + helm.sh/chart: rabbitmq-8.26.1 + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: rabbitmq + annotations: + checksum/config: d6c2caa9572f64a06d9f7daa34c664a186b4778cd1697ef8e59663152fc628f1 + checksum/secret: d764e7b3d999e7324d1afdfec6140092a612f04b6e0306818675815cec2f454f + spec: + + serviceAccountName: helm-rabbitmq + affinity: { } + securityContext: + fsGroup: 5001 + runAsUser: 5001 + terminationGracePeriodSeconds: 120 + initContainers: + - name: volume-permissions + image: 10.250.0.110:8033/cmii/bitnami-shell:10-debian-10-r140 + imagePullPolicy: "Always" + command: + - /bin/bash + args: + - -ec + - | + mkdir -p "/bitnami/rabbitmq/mnesia" + chown -R "5001:5001" "/bitnami/rabbitmq/mnesia" + securityContext: + runAsUser: 0 + resources: + limits: { } + requests: { } + volumeMounts: + - name: data + mountPath: /bitnami/rabbitmq/mnesia + containers: + - name: rabbitmq + image: 10.250.0.110:8033/cmii/rabbitmq:3.9.12-debian-10-r3 + imagePullPolicy: "Always" + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: MY_POD_NAMESPACE + valueFrom: + fieldRef: + fieldPath: metadata.namespace + - name: K8S_SERVICE_NAME + value: "helm-rabbitmq-headless" + - name: K8S_ADDRESS_TYPE + value: hostname + - name: RABBITMQ_FORCE_BOOT + value: "no" + - name: RABBITMQ_NODE_NAME + value: "rabbit@$(MY_POD_NAME).$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: K8S_HOSTNAME_SUFFIX + value: ".$(K8S_SERVICE_NAME).$(MY_POD_NAMESPACE).svc.cluster.local" + - name: RABBITMQ_MNESIA_DIR + value: "/bitnami/rabbitmq/mnesia/$(RABBITMQ_NODE_NAME)" + - name: RABBITMQ_LDAP_ENABLE + value: "no" + - name: RABBITMQ_LOGS + value: "-" + - name: RABBITMQ_ULIMIT_NOFILES + value: "65536" + - name: RABBITMQ_USE_LONGNAME + value: "true" + - name: RABBITMQ_ERL_COOKIE + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-erlang-cookie + - name: RABBITMQ_LOAD_DEFINITIONS + value: "no" + - name: RABBITMQ_SECURE_PASSWORD + value: "yes" + - name: RABBITMQ_USERNAME + value: "admin" + - name: RABBITMQ_PASSWORD + valueFrom: + secretKeyRef: + name: helm-rabbitmq + key: rabbitmq-password + - name: RABBITMQ_PLUGINS + value: "rabbitmq_management, rabbitmq_peer_discovery_k8s, rabbitmq_shovel, rabbitmq_shovel_management, rabbitmq_auth_backend_ldap" + ports: + - name: amqp + containerPort: 5672 + - name: dist + containerPort: 25672 + - name: dashboard + containerPort: 15672 + - name: epmd + containerPort: 4369 + livenessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q ping + initialDelaySeconds: 120 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 6 + readinessProbe: + exec: + command: + - /bin/bash + - -ec + - rabbitmq-diagnostics -q check_running && rabbitmq-diagnostics -q check_local_alarms + initialDelaySeconds: 10 + periodSeconds: 30 + timeoutSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + lifecycle: + preStop: + exec: + command: + - /bin/bash + - -ec + - | + if [[ -f /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh ]]; then + /opt/bitnami/scripts/rabbitmq/nodeshutdown.sh -t "120" -d "false" + else + rabbitmqctl stop_app + fi + resources: + limits: { } + requests: { } + volumeMounts: + - name: configuration + mountPath: /bitnami/rabbitmq/conf + - name: data + mountPath: /bitnami/rabbitmq/mnesia + volumes: + - name: configuration + configMap: + name: helm-rabbitmq-config + items: + - key: rabbitmq.conf + path: rabbitmq.conf + - name: data + persistentVolumeClaim: + claimName: helm-rabbitmq diff --git a/agent-operator/deploy/z_bjtg/k8s-redis.yaml b/agent-operator/deploy/z_bjtg/k8s-redis.yaml new file mode 100755 index 0000000..66ba7ba --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-redis.yaml @@ -0,0 +1,3492 @@ +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + +--- +apiVersion: v1 +kind: ServiceAccount +automountServiceAccountToken: true +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +--- +apiVersion: v1 +kind: Secret +metadata: + name: helm-redis + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +type: Opaque +data: + redis-password: "TWNhY2hlQDQ1MjI=" +--- +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-configuration + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + redis.conf: |- + # User-supplied common configuration: + # Enable AOF https://redis.io/topics/persistence#append-only-file + appendonly yes + # Disable RDB persistence, AOF persistence already enabled. + save "" + # End of common configuration + master.conf: |- + dir /data + # User-supplied master configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of master configuration + replica.conf: |- + dir /data + slave-read-only yes + # User-supplied replica configuration: + rename-command FLUSHDB "" + rename-command FLUSHALL "" + # End of replica configuration +--- +# Source: outside-deploy/charts/redis-db/templates/health-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-health + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + ping_readiness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_local.sh: |- + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -n "$REDIS_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h localhost \ + -p $REDIS_PORT \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ]; then + echo "$response" + exit 1 + fi + ping_liveness_master.sh: |- + #!/bin/bash + + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + [[ -n "$REDIS_MASTER_PASSWORD" ]] && export REDISCLI_AUTH="$REDIS_MASTER_PASSWORD" + response=$( + timeout -s 3 $1 \ + redis-cli \ + -h $REDIS_MASTER_HOST \ + -p $REDIS_MASTER_PORT_NUMBER \ + ping + ) + if [ "$response" != "PONG" ] && [ "$response" != "LOADING Redis is loading the dataset in memory" ]; then + echo "$response" + exit 1 + fi + ping_readiness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_readiness_local.sh" $1 || exit_status=$? + "$script_dir/ping_readiness_master.sh" $1 || exit_status=$? + exit $exit_status + ping_liveness_local_and_master.sh: |- + script_dir="$(dirname "$0")" + exit_status=0 + "$script_dir/ping_liveness_local.sh" $1 || exit_status=$? + "$script_dir/ping_liveness_master.sh" $1 || exit_status=$? + exit $exit_status +--- +# Source: outside-deploy/charts/redis-db/templates/scripts-configmap.yaml +apiVersion: v1 +kind: ConfigMap +metadata: + name: helm-redis-scripts + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +data: + start-master.sh: | + #!/bin/bash + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/master.conf ]];then + cp /opt/bitnami/redis/mounted-etc/master.conf /opt/bitnami/redis/etc/master.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/master.conf") + exec redis-server "${ARGS[@]}" + start-replica.sh: | + #!/bin/bash + + get_port() { + hostname="$1" + type="$2" + + port_var=$(echo "${hostname^^}_SERVICE_PORT_$type" | sed "s/-/_/g") + port=${!port_var} + + if [ -z "$port" ]; then + case $type in + "SENTINEL") + echo 26379 + ;; + "REDIS") + echo 6379 + ;; + esac + else + echo $port + fi + } + + get_full_hostname() { + hostname="$1" + echo "${hostname}.${HEADLESS_SERVICE}" + } + + REDISPORT=$(get_port "$HOSTNAME" "REDIS") + + [[ -f $REDIS_PASSWORD_FILE ]] && export REDIS_PASSWORD="$(< "${REDIS_PASSWORD_FILE}")" + [[ -f $REDIS_MASTER_PASSWORD_FILE ]] && export REDIS_MASTER_PASSWORD="$(< "${REDIS_MASTER_PASSWORD_FILE}")" + if [[ ! -f /opt/bitnami/redis/etc/replica.conf ]];then + cp /opt/bitnami/redis/mounted-etc/replica.conf /opt/bitnami/redis/etc/replica.conf + fi + if [[ ! -f /opt/bitnami/redis/etc/redis.conf ]];then + cp /opt/bitnami/redis/mounted-etc/redis.conf /opt/bitnami/redis/etc/redis.conf + fi + + echo "" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-port $REDISPORT" >> /opt/bitnami/redis/etc/replica.conf + echo "replica-announce-ip $(get_full_hostname "$HOSTNAME")" >> /opt/bitnami/redis/etc/replica.conf + ARGS=("--port" "${REDIS_PORT}") + ARGS+=("--slaveof" "${REDIS_MASTER_HOST}" "${REDIS_MASTER_PORT_NUMBER}") + ARGS+=("--requirepass" "${REDIS_PASSWORD}") + ARGS+=("--masterauth" "${REDIS_MASTER_PASSWORD}") + ARGS+=("--include" "/opt/bitnami/redis/etc/redis.conf") + ARGS+=("--include" "/opt/bitnami/redis/etc/replica.conf") + exec redis-server "${ARGS[@]}" +--- +# Source: outside-deploy/charts/redis-db/templates/headless-svc.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-headless + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus +spec: + type: ClusterIP + clusterIP: None + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg +--- +# Source: outside-deploy/charts/redis-db/templates/master/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + type: ClusterIP + + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/service.yaml +apiVersion: v1 +kind: Service +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + type: ClusterIP + ports: + - name: tcp-redis + port: 6379 + targetPort: redis + nodePort: null + selector: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica +--- +# Source: outside-deploy/charts/redis-db/templates/master/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-master + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + cmii.type: middleware + cmii.app: redis + app.kubernetes.io/component: master + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + affinity: { } + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-master.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: master + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + # One second longer than command timeout should prevent generation of zombie processes. + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc/ + - name: tmp + mountPath: /tmp + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: tmp + emptyDir: { } + - name: redis-data + emptyDir: { } +--- +# Source: outside-deploy/charts/redis-db/templates/replicas/statefulset.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: helm-redis-replicas + namespace: bjtg + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: redis-db + app.kubernetes.io/release: bjtg + app.kubernetes.io/component: replica + serviceName: helm-redis-headless + updateStrategy: + rollingUpdate: { } + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: redis-db + octopus.control: redis-db-wdd + app.kubernetes.io/release: bjtg + app.kubernetes.io/managed-by: octopus + app.kubernetes.io/component: replica + annotations: + checksum/configmap: b64aa5db67e6e63811f3c1095b9fce34d83c86a471fccdda0e48eedb53a179b0 + checksum/health: 6e0a6330e5ac63e565ae92af1444527d72d8897f91266f333555b3d323570623 + checksum/scripts: b88df93710b7c42a76006e20218f05c6e500e6cc2affd4bb1985832f03166e98 + checksum/secret: 43f1b0e20f9cb2de936bd182bc3683b720fc3cf4f4e76cb23c06a52398a50e8d + spec: + securityContext: + fsGroup: 1001 + serviceAccountName: helm-redis + terminationGracePeriodSeconds: 30 + containers: + - name: redis + image: 10.250.0.110:8033/cmii/redis:6.2.6-debian-10-r0 + imagePullPolicy: "Always" + securityContext: + runAsUser: 1001 + command: + - /bin/bash + args: + - -c + - /opt/bitnami/scripts/start-scripts/start-replica.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: REDIS_REPLICATION_MODE + value: slave + - name: REDIS_MASTER_HOST + value: helm-redis-master-0.helm-redis-headless.bjtg.svc.cluster.local + - name: REDIS_MASTER_PORT_NUMBER + value: "6379" + - name: ALLOW_EMPTY_PASSWORD + value: "no" + - name: REDIS_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_MASTER_PASSWORD + valueFrom: + secretKeyRef: + name: helm-redis + key: redis-password + - name: REDIS_TLS_ENABLED + value: "no" + - name: REDIS_PORT + value: "6379" + ports: + - name: redis + containerPort: 6379 + livenessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 6 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_liveness_local_and_master.sh 5 + readinessProbe: + initialDelaySeconds: 20 + periodSeconds: 5 + timeoutSeconds: 2 + successThreshold: 1 + failureThreshold: 5 + exec: + command: + - sh + - -c + - /health/ping_readiness_local_and_master.sh 1 + resources: + limits: + cpu: "2" + memory: 8Gi + requests: + cpu: "100m" + memory: 1Gi + volumeMounts: + - name: start-scripts + mountPath: /opt/bitnami/scripts/start-scripts + - name: health + mountPath: /health + - name: redis-data + mountPath: /data + subPath: + - name: config + mountPath: /opt/bitnami/redis/mounted-etc + - name: redis-tmp-conf + mountPath: /opt/bitnami/redis/etc + volumes: + - name: start-scripts + configMap: + name: helm-redis-scripts + defaultMode: 0755 + - name: health + configMap: + name: helm-redis-health + defaultMode: 0755 + - name: config + configMap: + name: helm-redis-configuration + - name: redis-tmp-conf + emptyDir: { } + - name: redis-data + emptyDir: { } + diff --git a/agent-operator/deploy/z_bjtg/k8s-srs.yaml b/agent-operator/deploy/z_bjtg/k8s-srs.yaml new file mode 100755 index 0000000..60a99df --- /dev/null +++ b/agent-operator/deploy/z_bjtg/k8s-srs.yaml @@ -0,0 +1,499 @@ +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: helm-live-srs-cm + namespace: bjtg + labels: + cmii.app: live-srs + cmii.type: live + octopus.control: wdd + app.kubernetes.io/managed-by: octopus + helm.sh/chart: cmlc-live-srs-rtc-2.0.0 +data: + srs.rtc.conf: |- + listen 30935; + max_connections 4096; + srs_log_tank console; + srs_log_level info; + srs_log_file /home/srs.log; + daemon off; + http_api { + enabled on; + listen 1985; + crossdomain on; + } + stats { + network 0; + } + http_server { + enabled on; + listen 8080; + dir /home/hls; + } + srt_server { + enabled on; + listen 30556; + maxbw 1000000000; + connect_timeout 4000; + peerlatency 600; + recvlatency 600; + } + rtc_server { + enabled on; + listen 30090; + candidate $CANDIDATE; + } + vhost __defaultVhost__ { + http_hooks { + enabled on; + on_publish http://helm-live-op-svc-v2:8080/hooks/on_push; + } + http_remux { + enabled on; + } + rtc { + enabled on; + rtmp_to_rtc on; + rtc_to_rtmp on; + keep_bframe off; + } + tcp_nodelay on; + min_latency on; + play { + gop_cache off; + mw_latency 100; + mw_msgs 10; + } + publish { + firstpkt_timeout 8000; + normal_timeout 4000; + mr on; + } + dvr { + enabled off; + dvr_path /home/dvr/[app]/[stream]/[2006][01]/[timestamp].mp4; + dvr_plan session; + } + hls { + enabled on; + hls_path /home/hls; + hls_fragment 10; + hls_window 60; + hls_m3u8_file [app]/[stream].m3u8; + hls_ts_file [app]/[stream]/[2006][01][02]/[timestamp]-[duration].ts; + hls_cleanup on; + hls_entry_prefix http://10.250.0.110:8888; + } + } +--- +kind: Service +apiVersion: v1 +metadata: + name: helm-live-srs-svc-exporter + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus +spec: + ports: + - name: rtmp + protocol: TCP + port: 30935 + targetPort: 30935 + nodePort: 30935 + - name: rtc + protocol: UDP + port: 30090 + targetPort: 30090 + nodePort: 30090 + - name: rtc-tcp + protocol: TCP + port: 30090 + targetPort: 30090 + nodePort: 30090 + - name: srt + protocol: UDP + port: 30556 + targetPort: 30556 + nodePort: 30556 + - name: api + protocol: TCP + port: 1985 + targetPort: 1985 + nodePort: 30557 + selector: + srs-role: rtc + type: NodePort + sessionAffinity: None + externalTrafficPolicy: Cluster + +--- +kind: Service +apiVersion: v1 +metadata: + name: helm-live-srs-svc + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus +spec: + ports: + - name: http + protocol: TCP + port: 8080 + targetPort: 8080 + - name: api + protocol: TCP + port: 1985 + targetPort: 1985 + selector: + srs-role: rtc + type: ClusterIP + sessionAffinity: None + +--- +kind: Service +apiVersion: v1 +metadata: + name: helm-live-srsrtc-svc + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus +spec: + ports: + - name: rtmp + protocol: TCP + port: 30935 + targetPort: 30935 + selector: + srs-role: rtc + type: ClusterIP + sessionAffinity: None + +--- +kind: StatefulSet +apiVersion: apps/v1 +metadata: + name: helm-live-srs-rtc + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus + cmii.app: live-srs + cmii.type: live + helm.sh/chart: cmlc-live-srs-rtc-2.0.0 + srs-role: rtc +spec: + replicas: 1 + selector: + matchLabels: + srs-role: rtc + template: + metadata: + creationTimestamp: null + labels: + srs-role: rtc + spec: + volumes: + - name: srs-conf-file + configMap: + name: helm-live-srs-cm + items: + - key: srs.rtc.conf + path: docker.conf + defaultMode: 420 + - name: srs-vol + emptyDir: + sizeLimit: 8Gi + containers: + - name: srs-rtc + image: 10.250.0.110:8033/cmii/srs:v5.0.195 + ports: + - name: srs-rtmp + containerPort: 30935 + protocol: TCP + - name: srs-api + containerPort: 1985 + protocol: TCP + - name: srs-flv + containerPort: 8080 + protocol: TCP + - name: srs-webrtc + containerPort: 30090 + protocol: UDP + - name: srs-webrtc-tcp + containerPort: 30090 + protocol: TCP + - name: srs-srt + containerPort: 30556 + protocol: UDP + env: + - name: CANDIDATE + value: 10.250.0.110 + resources: + limits: + cpu: 1200m + memory: 6Gi + requests: + cpu: 100m + memory: 256Mi + volumeMounts: + - name: srs-conf-file + mountPath: /usr/local/srs/conf/docker.conf + subPath: docker.conf + - name: srs-vol + mountPath: /home/dvr + subPath: bjtg/helm-live/dvr + - name: srs-vol + mountPath: /home/hls + subPath: bjtg/helm-live/hls + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + imagePullPolicy: Always + - name: oss-adaptor + image: 10.250.0.110:8033/cmii/cmii-srs-oss-adaptor:2023-SA + env: + - name: OSS_ENDPOINT + value: 'http://10.250.0.110:9000' + - name: OSS_AK + value: cmii + - name: OSS_SK + value: 'B#923fC7mk' + - name: OSS_BUCKET + value: live-cluster-hls + - name: SRS_OP + value: 'http://helm-live-op-svc-v2:8080' + - name: MYSQL_ENDPOINT + value: 'helm-mysql:3306' + - name: MYSQL_USERNAME + value: k8s_admin + - name: MYSQL_PASSWORD + value: fP#UaH6qQ3)8 + - name: MYSQL_DATABASE + value: cmii_live_srs_op + - name: MYSQL_TABLE + value: live_segment + - name: LOG_LEVEL + value: info + - name: OSS_META + value: 'yes' + resources: + limits: + cpu: 1200m + memory: 4Gi + requests: + cpu: 100m + memory: 256Mi + volumeMounts: + - name: srs-vol + mountPath: /cmii/share/hls + subPath: bjtg/helm-live/hls + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + imagePullPolicy: Always + restartPolicy: Always + terminationGracePeriodSeconds: 30 + dnsPolicy: ClusterFirst + securityContext: { } + imagePullSecrets: + - name: harborsecret + affinity: { } + schedulerName: default-scheduler + serviceName: helm-live-srsrtc-svc + podManagementPolicy: OrderedReady + updateStrategy: + type: RollingUpdate + rollingUpdate: + partition: 0 + revisionHistoryLimit: 10 +--- +# live-srs部分 +--- +kind: Deployment +apiVersion: apps/v1 +metadata: + name: helm-live-op-v2 + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus + cmii.app: live-engine + cmii.type: live + helm.sh/chart: cmlc-live-live-op-2.0.0 + live-role: op-v2 +spec: + replicas: 1 + selector: + matchLabels: + live-role: op-v2 + template: + metadata: + creationTimestamp: null + labels: + live-role: op-v2 + spec: + volumes: + - name: srs-conf-file + configMap: + name: helm-live-op-cm-v2 + items: + - key: live.op.conf + path: bootstrap.yaml + defaultMode: 420 + containers: + - name: helm-live-op-v2 + image: 10.250.0.110:8033/cmii/cmii-live-operator:5.2.0 + ports: + - name: operator + containerPort: 8080 + protocol: TCP + resources: + limits: + cpu: 4800m + memory: 4Gi + requests: + cpu: 100m + memory: 256Mi + volumeMounts: + - name: srs-conf-file + mountPath: /cmii/bootstrap.yaml + subPath: bootstrap.yaml + livenessProbe: + httpGet: + path: /cmii/ping + port: 8080 + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + readinessProbe: + httpGet: + path: /cmii/ping + port: 8080 + scheme: HTTP + initialDelaySeconds: 60 + timeoutSeconds: 5 + periodSeconds: 20 + successThreshold: 1 + failureThreshold: 3 + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + imagePullPolicy: Always + restartPolicy: Always + terminationGracePeriodSeconds: 30 + dnsPolicy: ClusterFirst + securityContext: { } + imagePullSecrets: + - name: harborsecret + affinity: { } + schedulerName: default-scheduler + strategy: + type: RollingUpdate + rollingUpdate: + maxUnavailable: 25% + maxSurge: 25% + revisionHistoryLimit: 10 + progressDeadlineSeconds: 600 +--- +kind: Service +apiVersion: v1 +metadata: + name: helm-live-op-svc-v2 + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus +spec: + ports: + - protocol: TCP + port: 8080 + targetPort: 8080 + nodePort: 30333 + selector: + live-role: op-v2 + type: NodePort + sessionAffinity: None +--- +kind: Service +apiVersion: v1 +metadata: + name: helm-live-op-svc + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus +spec: + ports: + - protocol: TCP + port: 8080 + targetPort: 8080 + selector: + live-role: op + type: ClusterIP + sessionAffinity: None +--- +kind: ConfigMap +apiVersion: v1 +metadata: + name: helm-live-op-cm-v2 + namespace: bjtg + labels: + octopus.control: wdd + app.kubernetes.io/managed-by: octopus + cmii.app: live-engine + cmii.type: live +data: + live.op.conf: |- + server: + port: 8080 + spring: + main: + allow-bean-definition-overriding: true + allow-circular-references: true + application: + name: cmii-live-operator + platform: + info: + name: cmii-live-operator + description: cmii-live-operator + version: 5.5.0 + scanPackage: com.cmii.live.op + cloud: + nacos: + config: + username: developer + password: N@cos14Good + server-addr: helm-nacos:8848 + extension-configs: + - data-id: cmii-live-operator.yml + group: 5.5.0 + refresh: true + shared-configs: + - data-id: cmii-backend-system.yml + group: 5.5.0 + refresh: true + discovery: + enabled: false + + live: + engine: + type: srs + endpoint: 'http://helm-live-srs-svc:1985' + + proto: + rtmp: 'rtmp://10.250.0.110:30935' + rtsp: 'rtsp://10.250.0.110:30554' + srt: 'srt://10.250.0.110:30556' + flv: 'http://10.250.0.110:30500' + hls: 'http://10.250.0.110:30500' + rtc: 'webrtc://10.250.0.110:30557' + replay: 'https://10.250.0.110:30333' + minio: + endpoint: http://10.250.0.110:9000 + access-key: cmii + secret-key: B#923fC7mk + bucket: live-cluster-hls diff --git a/agent-operator/go.mod b/agent-operator/go.mod old mode 100644 new mode 100755 diff --git a/agent-operator/go.sum b/agent-operator/go.sum old mode 100644 new mode 100755 diff --git a/agent-operator/image/CmiiImageOperator.go b/agent-operator/image/CmiiImageOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/image/CmiiImageOperator_test.go b/agent-operator/image/CmiiImageOperator_test.go old mode 100644 new mode 100755 diff --git a/agent-operator/image/HarborOperator.go b/agent-operator/image/HarborOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/image/HarborOperator_test.go b/agent-operator/image/HarborOperator_test.go old mode 100644 new mode 100755 diff --git a/agent-operator/image/image_upload_oss.sh b/agent-operator/image/image_upload_oss.sh old mode 100644 new mode 100755 diff --git a/agent-operator/image/middle-image.txt b/agent-operator/image/middle-image.txt old mode 100644 new mode 100755 diff --git a/agent-operator/image/octopus_image_sync.sh b/agent-operator/image/octopus_image_sync.sh old mode 100644 new mode 100755 diff --git a/agent-operator/image/octopus_image_update.sh b/agent-operator/image/octopus_image_update.sh old mode 100644 new mode 100755 diff --git a/agent-operator/image/rke-image.txt b/agent-operator/image/rke-image.txt old mode 100644 new mode 100755 diff --git a/agent-operator/log/cmii-update-log.txt b/agent-operator/log/cmii-update-log.txt old mode 100644 new mode 100755 index 9aef43e..dabdb10 --- a/agent-operator/log/cmii-update-log.txt +++ b/agent-operator/log/cmii-update-log.txt @@ -55,3 +55,4 @@ 2024-06-12-13-40-00 uavcloud-demo cmii-uav-industrial-portfolio 5.6.0-061101 5.6.0-061201 2024-06-12-13-45-01 uavcloud-demo cmii-uav-industrial-portfolio 5.6.0-061201 5.6.0-061101 2024-06-12-13-47-35 uavcloud-demo cmii-uav-industrial-portfolio 5.6.0-061101 5.6.0-061201 +2024-06-13-17-23-00 uavcloud-demo cmii-uav-industrial-portfolio 5.6.0-061202 5.6.0-061302 diff --git a/agent-operator/main.go b/agent-operator/main.go old mode 100644 new mode 100755 diff --git a/agent-operator/mysql/MysqlOperator.go b/agent-operator/mysql/MysqlOperator.go old mode 100644 new mode 100755 diff --git a/agent-operator/mysql/cmii_mysql_dump.sh b/agent-operator/mysql/cmii_mysql_dump.sh old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/bgtg/Config.go b/agent-operator/real_project/bgtg/Config.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/0-install_all_demand.sh b/agent-operator/real_project/cqga/0-install_all_demand.sh old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/1.mount_volume.sh b/agent-operator/real_project/cqga/1.mount_volume.sh old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/2-harbor-docker-compose.yaml b/agent-operator/real_project/cqga/2-harbor-docker-compose.yaml old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/3-rke-cluster.yml b/agent-operator/real_project/cqga/3-rke-cluster.yml old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/operator.go b/agent-operator/real_project/cqga/operator.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/cqga/project-info.txt b/agent-operator/real_project/cqga/project-info.txt old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/octopus-agent-run.txt b/agent-operator/real_project/octopus-agent-run.txt old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/proxy_project/linux/port_linux_amd64 b/agent-operator/real_project/proxy_project/linux/port_linux_amd64 old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/proxy_project/linux/socks5_linux_amd64 b/agent-operator/real_project/proxy_project/linux/socks5_linux_amd64 old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/proxy_project/windows/port_win64.exe b/agent-operator/real_project/proxy_project/windows/port_win64.exe old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/proxy_project/windows/socks5_win64.exe b/agent-operator/real_project/proxy_project/windows/socks5_win64.exe old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/szga/Config.go b/agent-operator/real_project/szga/Config.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/xmyd/CmiiConfig.go b/agent-operator/real_project/xmyd/CmiiConfig.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/xzyd/Config.go b/agent-operator/real_project/xzyd/Config.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/zjjt/Config.go b/agent-operator/real_project/zjjt/Config.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/zjjt/RealImageConfig.go b/agent-operator/real_project/zjjt/RealImageConfig.go old mode 100644 new mode 100755 diff --git a/agent-operator/real_project/zyga/operator.go b/agent-operator/real_project/zyga/operator.go old mode 100644 new mode 100755 diff --git a/agent-operator/tools/JsonMessage.go b/agent-operator/tools/JsonMessage.go old mode 100644 new mode 100755 diff --git a/agent/.gitignore b/agent/.gitignore old mode 100644 new mode 100755 diff --git a/agent/Dockerfile b/agent/Dockerfile old mode 100644 new mode 100755 diff --git a/agent/Dockerfile-wsl2 b/agent/Dockerfile-wsl2 old mode 100644 new mode 100755 diff --git a/agent/EasyCodeConfig.json b/agent/EasyCodeConfig.json old mode 100644 new mode 100755 diff --git a/agent/agent-entrypoint.sh b/agent/agent-entrypoint.sh old mode 100644 new mode 100755 diff --git a/agent/all-agent-hosts.txt b/agent/all-agent-hosts.txt old mode 100644 new mode 100755 diff --git a/agent/current-env.txt b/agent/current-env.txt old mode 100644 new mode 100755 diff --git a/agent/pom.xml b/agent/pom.xml old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/AgentApplication.java b/agent/src/main/java/io/wdd/agent/AgentApplication.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/agent/AgentOperationInfoService.java b/agent/src/main/java/io/wdd/agent/agent/AgentOperationInfoService.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/agent/AgentRebootUpdateService.java b/agent/src/main/java/io/wdd/agent/agent/AgentRebootUpdateService.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/beans/executor/CommandLog.java b/agent/src/main/java/io/wdd/agent/config/beans/executor/CommandLog.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/beans/executor/StreamSenderEntity.java b/agent/src/main/java/io/wdd/agent/config/beans/executor/StreamSenderEntity.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/beans/init/AgentServerInfo.java b/agent/src/main/java/io/wdd/agent/config/beans/init/AgentServerInfo.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/OctopusRabbitMQAdminConfig.java b/agent/src/main/java/io/wdd/agent/config/message/OctopusRabbitMQAdminConfig.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/AbstractOctopusMessageHandler.java b/agent/src/main/java/io/wdd/agent/config/message/handler/AbstractOctopusMessageHandler.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerAgent.java b/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerAgent.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerBlackHole.java b/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerBlackHole.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerExecutor.java b/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerExecutor.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerInit.java b/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerInit.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerStatus.java b/agent/src/main/java/io/wdd/agent/config/message/handler/OMHandlerStatus.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/utils/AgentCommonThreadPool.java b/agent/src/main/java/io/wdd/agent/config/utils/AgentCommonThreadPool.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/utils/BeanFactoryUtils.java b/agent/src/main/java/io/wdd/agent/config/utils/BeanFactoryUtils.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/utils/OctopusObjectMapper.java b/agent/src/main/java/io/wdd/agent/config/utils/OctopusObjectMapper.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/config/wsl2-fixed-ip.bat b/agent/src/main/java/io/wdd/agent/config/wsl2-fixed-ip.bat old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/CommandExecutor.java b/agent/src/main/java/io/wdd/agent/executor/CommandExecutor.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/FunctionExecutor.java b/agent/src/main/java/io/wdd/agent/executor/FunctionExecutor.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/config/CommandPipelineBuilder.java b/agent/src/main/java/io/wdd/agent/executor/config/CommandPipelineBuilder.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/config/ExecutorFunctionNacosCollector.java b/agent/src/main/java/io/wdd/agent/executor/config/ExecutorFunctionNacosCollector.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/config/GetClassPath.java b/agent/src/main/java/io/wdd/agent/executor/config/GetClassPath.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/config/RedisConfiguration.java b/agent/src/main/java/io/wdd/agent/executor/config/RedisConfiguration.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/reply/SimpleStreamSender.java b/agent/src/main/java/io/wdd/agent/executor/reply/SimpleStreamSender.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/reply/StreamSender.java b/agent/src/main/java/io/wdd/agent/executor/reply/StreamSender.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/status/AppStatusExecutor.java b/agent/src/main/java/io/wdd/agent/executor/status/AppStatusExecutor.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/status/NetworkInterfaceSpeedCallable.java b/agent/src/main/java/io/wdd/agent/executor/status/NetworkInterfaceSpeedCallable.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/status/NetworkStatusExecutor.java b/agent/src/main/java/io/wdd/agent/executor/status/NetworkStatusExecutor.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/status/SingleAppStatusCallable.java b/agent/src/main/java/io/wdd/agent/executor/status/SingleAppStatusCallable.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/thread/CommandExecLogCache.java b/agent/src/main/java/io/wdd/agent/executor/thread/CommandExecLogCache.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/thread/LogToStreamSender.java b/agent/src/main/java/io/wdd/agent/executor/thread/LogToStreamSender.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/thread/LogToSysOut.java b/agent/src/main/java/io/wdd/agent/executor/thread/LogToSysOut.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/executor/web/TestCommandExecutorController.java b/agent/src/main/java/io/wdd/agent/executor/web/TestCommandExecutorController.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/CollectSystemInfo.java b/agent/src/main/java/io/wdd/agent/initialization/bootup/CollectSystemInfo.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/OctopusAgentInitService.java b/agent/src/main/java/io/wdd/agent/initialization/bootup/OctopusAgentInitService.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/AutoDeployOctopusAgent.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/AutoDeployOctopusAgent.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/BootUpAgent.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/BootUpAgent.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/MaualRunAgent.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/deprecated/MaualRunAgent.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/bootup.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/bootup.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/linux-init-LapPro.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/linux-init-LapPro.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/teddysun-benchmark-linux-info.png b/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/teddysun-benchmark-linux-info.png old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/teddysun-benchmark.sh b/agent/src/main/java/io/wdd/agent/initialization/bootup/reference/teddysun-benchmark.sh old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/message/GenOctopusRabbitMQConnection.java b/agent/src/main/java/io/wdd/agent/initialization/message/GenOctopusRabbitMQConnection.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/message/InitRabbitMQConnector.java b/agent/src/main/java/io/wdd/agent/initialization/message/InitRabbitMQConnector.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/initialization/webtest/SendServerInfoController.java b/agent/src/main/java/io/wdd/agent/initialization/webtest/SendServerInfoController.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/message/OMessageToServerSender.java b/agent/src/main/java/io/wdd/agent/message/OMessageToServerSender.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/message/handler/OctopusMessageHandler.java b/agent/src/main/java/io/wdd/agent/message/handler/OctopusMessageHandler.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/status/AgentStatusCollector.java b/agent/src/main/java/io/wdd/agent/status/AgentStatusCollector.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/status/HealthyReporter.java b/agent/src/main/java/io/wdd/agent/status/HealthyReporter.java old mode 100644 new mode 100755 diff --git a/agent/src/main/java/io/wdd/agent/todo.md b/agent/src/main/java/io/wdd/agent/todo.md old mode 100644 new mode 100755 diff --git a/agent/src/main/resources/application-local.yaml b/agent/src/main/resources/application-local.yaml old mode 100644 new mode 100755 diff --git a/agent/src/main/resources/bootstrap.yml b/agent/src/main/resources/bootstrap.yml old mode 100644 new mode 100755 diff --git a/agent/src/main/resources/nginx-loadbalancer.conf b/agent/src/main/resources/nginx-loadbalancer.conf old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/AgentApplicationTests.java b/agent/src/test/java/io/wdd/agent/AgentApplicationTests.java old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/AppStatusCheckTest.java b/agent/src/test/java/io/wdd/agent/AppStatusCheckTest.java old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/CommandTest.java b/agent/src/test/java/io/wdd/agent/CommandTest.java old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/OSHITest.java b/agent/src/test/java/io/wdd/agent/OSHITest.java old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/SimpleTest.java b/agent/src/test/java/io/wdd/agent/SimpleTest.java old mode 100644 new mode 100755 diff --git a/agent/src/test/java/io/wdd/agent/test.sh b/agent/src/test/java/io/wdd/agent/test.sh old mode 100644 new mode 100755 diff --git a/common/.gitignore b/common/.gitignore old mode 100644 new mode 100755 diff --git a/common/pom.xml b/common/pom.xml old mode 100644 new mode 100755 diff --git a/common/src/main/java/io/wdd/common/CommonApplication.java b/common/src/main/java/io/wdd/common/CommonApplication.java old mode 100644 new mode 100755 diff --git a/common/src/main/java/io/wdd/common/beans/executor/ExecutorFunctionMessage.java b/common/src/main/java/io/wdd/common/beans/executor/ExecutorFunctionMessage.java old mode 100644 new mode 100755 diff --git a/common/src/main/resources/application.properties b/common/src/main/resources/application.properties old mode 100644 new mode 100755 diff --git a/common/src/test/java/io/wdd/common/CommonApplicationTests.java b/common/src/test/java/io/wdd/common/CommonApplicationTests.java old mode 100644 new mode 100755 diff --git a/index.html b/index.html old mode 100644 new mode 100755 diff --git a/message_pusher/Dockerfile b/message_pusher/Dockerfile old mode 100644 new mode 100755 diff --git a/message_pusher/cmii/CmiiMessage.go b/message_pusher/cmii/CmiiMessage.go old mode 100644 new mode 100755 diff --git a/message_pusher/cmii/CmiiMessage_test.go b/message_pusher/cmii/CmiiMessage_test.go old mode 100644 new mode 100755 diff --git a/message_pusher/cmii/ImageSyncMessage.go b/message_pusher/cmii/ImageSyncMessage.go old mode 100644 new mode 100755 diff --git a/message_pusher/docs/docs.go b/message_pusher/docs/docs.go old mode 100644 new mode 100755 diff --git a/message_pusher/docs/swagger.json b/message_pusher/docs/swagger.json old mode 100644 new mode 100755 diff --git a/message_pusher/docs/swagger.yaml b/message_pusher/docs/swagger.yaml old mode 100644 new mode 100755 diff --git a/message_pusher/go.mod b/message_pusher/go.mod old mode 100644 new mode 100755 diff --git a/message_pusher/go.sum b/message_pusher/go.sum old mode 100644 new mode 100755 diff --git a/message_pusher/main.go b/message_pusher/main.go old mode 100644 new mode 100755 diff --git a/message_pusher/pusher/client.go b/message_pusher/pusher/client.go old mode 100644 new mode 100755 diff --git a/message_pusher/pusher/config.go b/message_pusher/pusher/config.go old mode 100644 new mode 100755 diff --git a/message_pusher/pusher/options.go b/message_pusher/pusher/options.go old mode 100644 new mode 100755 diff --git a/message_pusher/router/CmiiRouter.go b/message_pusher/router/CmiiRouter.go old mode 100644 new mode 100755 diff --git a/message_pusher/router/ImageSyncRouter.go b/message_pusher/router/ImageSyncRouter.go old mode 100644 new mode 100755 diff --git a/message_pusher/router/OctopusRouter.go b/message_pusher/router/OctopusRouter.go old mode 100644 new mode 100755 diff --git a/message_pusher/swagger.ps1 b/message_pusher/swagger.ps1 old mode 100644 new mode 100755 diff --git a/pom.xml b/pom.xml old mode 100644 new mode 100755 diff --git a/port_forwarding/CHANGELOG b/port_forwarding/CHANGELOG old mode 100644 new mode 100755 diff --git a/port_forwarding/Images/mutil_forward.png b/port_forwarding/Images/mutil_forward.png old mode 100644 new mode 100755 diff --git a/port_forwarding/Images/portforward_framework.png b/port_forwarding/Images/portforward_framework.png old mode 100644 new mode 100755 diff --git a/port_forwarding/Images/restricted_forward.png b/port_forwarding/Images/restricted_forward.png old mode 100644 new mode 100755 diff --git a/port_forwarding/Images/simple_forward.png b/port_forwarding/Images/simple_forward.png old mode 100644 new mode 100755 diff --git a/port_forwarding/LICENSE b/port_forwarding/LICENSE old mode 100644 new mode 100755 diff --git a/port_forwarding/README.md b/port_forwarding/README.md old mode 100644 new mode 100755 diff --git a/port_forwarding/build.sh b/port_forwarding/build.sh old mode 100644 new mode 100755 diff --git a/port_forwarding/forward.go b/port_forwarding/forward.go old mode 100644 new mode 100755 diff --git a/port_forwarding/go.mod b/port_forwarding/go.mod old mode 100644 new mode 100755 diff --git a/port_forwarding/log.go b/port_forwarding/log.go old mode 100644 new mode 100755 diff --git a/port_forwarding/main.go b/port_forwarding/main.go old mode 100644 new mode 100755 diff --git a/port_forwarding/script/1-proxy-setup.ps1 b/port_forwarding/script/1-proxy-setup.ps1 old mode 100644 new mode 100755 diff --git a/port_forwarding/script/2-代理全部cmii端口.ps1 b/port_forwarding/script/2-代理全部cmii端口.ps1 old mode 100644 new mode 100755 diff --git a/port_forwarding/script/start-proxy.sh b/port_forwarding/script/start-proxy.sh old mode 100644 new mode 100755 diff --git a/port_forwarding/tcp.go b/port_forwarding/tcp.go old mode 100644 new mode 100755 diff --git a/port_forwarding/tcp_learn/socket_learn.go b/port_forwarding/tcp_learn/socket_learn.go old mode 100644 new mode 100755 diff --git a/port_forwarding/tcp_learn/socket_learn_test.go b/port_forwarding/tcp_learn/socket_learn_test.go old mode 100644 new mode 100755 diff --git a/port_forwarding/udp.go b/port_forwarding/udp.go old mode 100644 new mode 100755 diff --git a/server-go/entity/ServerInfo.go b/server-go/entity/ServerInfo.go old mode 100644 new mode 100755 diff --git a/server-go/entity/global.go b/server-go/entity/global.go old mode 100644 new mode 100755 diff --git a/server-go/go.mod b/server-go/go.mod old mode 100644 new mode 100755 diff --git a/server-go/go.sum b/server-go/go.sum old mode 100644 new mode 100755 diff --git a/server-go/main.go b/server-go/main.go old mode 100644 new mode 100755 diff --git a/server-go/router/ServerRouter.go b/server-go/router/ServerRouter.go old mode 100644 new mode 100755 diff --git a/server-go/router/router.go b/server-go/router/router.go old mode 100644 new mode 100755 diff --git a/server-go/service/DBConfig.go b/server-go/service/DBConfig.go old mode 100644 new mode 100755 diff --git a/server-go/service/ServerAPI.go b/server-go/service/ServerAPI.go old mode 100644 new mode 100755 diff --git a/server-go/test/channelTest.go b/server-go/test/channelTest.go old mode 100644 new mode 100755 diff --git a/server/.gitignore b/server/.gitignore old mode 100644 new mode 100755 diff --git a/server/Dockerfile b/server/Dockerfile old mode 100644 new mode 100755 diff --git a/server/Dockerfile-shanghai-remote-idea b/server/Dockerfile-shanghai-remote-idea old mode 100644 new mode 100755 diff --git a/server/Dockerfile-teamcity b/server/Dockerfile-teamcity old mode 100644 new mode 100755 diff --git a/server/pom.xml b/server/pom.xml old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/ServerApplication.java b/server/src/main/java/io/wdd/ServerApplication.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/config/CorsConfig.java b/server/src/main/java/io/wdd/common/config/CorsConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/config/JedisConfig.java b/server/src/main/java/io/wdd/common/config/JedisConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/config/OctopusObjectMapperConfig.java b/server/src/main/java/io/wdd/common/config/OctopusObjectMapperConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/handler/GlobalExceptionHandler.java b/server/src/main/java/io/wdd/common/handler/GlobalExceptionHandler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/handler/MyMessageSource.java b/server/src/main/java/io/wdd/common/handler/MyMessageSource.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/handler/MyRuntimeException.java b/server/src/main/java/io/wdd/common/handler/MyRuntimeException.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/response/R.java b/server/src/main/java/io/wdd/common/response/R.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/response/ResultStat.java b/server/src/main/java/io/wdd/common/response/ResultStat.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/DataUnit.java b/server/src/main/java/io/wdd/common/utils/DataUnit.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/FakeDataUtils.java b/server/src/main/java/io/wdd/common/utils/FakeDataUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/FormatUtils.java b/server/src/main/java/io/wdd/common/utils/FormatUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/FunctionReader.java b/server/src/main/java/io/wdd/common/utils/FunctionReader.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/MessageUtils.java b/server/src/main/java/io/wdd/common/utils/MessageUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/OctopusRabbitTemplateConfig.java b/server/src/main/java/io/wdd/common/utils/OctopusRabbitTemplateConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/common/utils/TimeUtils.java b/server/src/main/java/io/wdd/common/utils/TimeUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/AppFunctionEnum.java b/server/src/main/java/io/wdd/func/auto/beans/AppFunctionEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/BaseFunctionEnum.java b/server/src/main/java/io/wdd/func/auto/beans/BaseFunctionEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/HarborFunctionEnum.java b/server/src/main/java/io/wdd/func/auto/beans/HarborFunctionEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/ImageFunctionEnum.java b/server/src/main/java/io/wdd/func/auto/beans/ImageFunctionEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/ImageSyncContext.java b/server/src/main/java/io/wdd/func/auto/beans/ImageSyncContext.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/beans/ProjectDeployContext.java b/server/src/main/java/io/wdd/func/auto/beans/ProjectDeployContext.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/AppFuncScheduler.java b/server/src/main/java/io/wdd/func/auto/service/AppFuncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/BaseFuncScheduler.java b/server/src/main/java/io/wdd/func/auto/service/BaseFuncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/FuncService.java b/server/src/main/java/io/wdd/func/auto/service/FuncService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/FuncServiceImpl.java b/server/src/main/java/io/wdd/func/auto/service/FuncServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/HarborFuncScheduler.java b/server/src/main/java/io/wdd/func/auto/service/HarborFuncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/auto/service/ImageFuncScheduler.java b/server/src/main/java/io/wdd/func/auto/service/ImageFuncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/controller/AutoController.java b/server/src/main/java/io/wdd/func/controller/AutoController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/controller/OSSController.java b/server/src/main/java/io/wdd/func/controller/OSSController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/controller/XrayController.java b/server/src/main/java/io/wdd/func/controller/XrayController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/config/OSSConfiguration.java b/server/src/main/java/io/wdd/func/oss/config/OSSConfiguration.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/config/OctopusObjectSummary.java b/server/src/main/java/io/wdd/func/oss/config/OctopusObjectSummary.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/config/OssConfig.java b/server/src/main/java/io/wdd/func/oss/config/OssConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/service/OSSCoreService.java b/server/src/main/java/io/wdd/func/oss/service/OSSCoreService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/service/OSSCoreServiceImpl.java b/server/src/main/java/io/wdd/func/oss/service/OSSCoreServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/oss/service/OssBackendSelect.java b/server/src/main/java/io/wdd/func/oss/service/OssBackendSelect.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessHTTPTemplateClass.java b/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessHTTPTemplateClass.java old mode 100644 new mode 100755 index 4ea5127..573118f --- a/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessHTTPTemplateClass.java +++ b/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessHTTPTemplateClass.java @@ -1,10 +1,10 @@ package io.wdd.func.xray.beans.config; +import io.wdd.func.xray.beans.xray.protocol.inbound.vmess.VMESS; import io.wdd.func.xray.beans.xray.transport.HttpHeaderObject; import io.wdd.func.xray.beans.xray.transport.InboundObject; import io.wdd.func.xray.beans.xray.transport.StreamSettingsObject; import io.wdd.func.xray.beans.xray.transport.TcpObject; -import io.wdd.func.xray.beans.xray.protocol.inbound.vmess.VMESS; import static io.wdd.func.xray.beans.config.TcpHttpHeaderTemplate.HttpResponseTemplate; diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessTCPTemplateClass.java b/server/src/main/java/io/wdd/func/xray/beans/config/InboundVmessTCPTemplateClass.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/LogTemplateClass.java b/server/src/main/java/io/wdd/func/xray/beans/config/LogTemplateClass.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/OutboundVmessHTTPTemplateClass.java b/server/src/main/java/io/wdd/func/xray/beans/config/OutboundVmessHTTPTemplateClass.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/OutboundVmessTCPTemplateClass.java b/server/src/main/java/io/wdd/func/xray/beans/config/OutboundVmessTCPTemplateClass.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/config/TcpHttpHeaderTemplate.java b/server/src/main/java/io/wdd/func/xray/beans/config/TcpHttpHeaderTemplate.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/node/Edge.java b/server/src/main/java/io/wdd/func/xray/beans/node/Edge.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/node/Graph.java b/server/src/main/java/io/wdd/func/xray/beans/node/Graph.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/node/ProxyNode.java b/server/src/main/java/io/wdd/func/xray/beans/node/ProxyNode.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/node/ProxyNodeType.java b/server/src/main/java/io/wdd/func/xray/beans/node/ProxyNodeType.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/node/XrayConfigInfo.java b/server/src/main/java/io/wdd/func/xray/beans/node/XrayConfigInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/Account.java b/server/src/main/java/io/wdd/func/xray/beans/xray/Account.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/ApiObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/ApiObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/Client.java b/server/src/main/java/io/wdd/func/xray/beans/xray/Client.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/DnsObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/DnsObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/LogObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/LogObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/Order.java b/server/src/main/java/io/wdd/func/xray/beans/xray/Order.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/PolicyObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/PolicyObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/ReverseObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/ReverseObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/RoutingObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/RoutingObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/RuleObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/RuleObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/StatsObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/StatsObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/User.java b/server/src/main/java/io/wdd/func/xray/beans/xray/User.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/XrayConfig.java b/server/src/main/java/io/wdd/func/xray/beans/xray/XrayConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/DokodemoDoor.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/DokodemoDoor.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/ClientObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/ClientObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/FallbackObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/FallbackObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/VLESS.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/VLESS.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/VlessConfig.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vless/VlessConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/ClientObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/ClientObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/DefaultObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/DefaultObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/DetourObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/DetourObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/InboundConfigurationObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/InboundConfigurationObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/VMESS.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/inbound/vmess/VMESS.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/Blackhole.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/Blackhole.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/DNS.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/DNS.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/Freedom.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/Freedom.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/OutboundConfigurationObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/OutboundConfigurationObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VLESS.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VLESS.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VMESS.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VMESS.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VmessOutSettings.java b/server/src/main/java/io/wdd/func/xray/beans/xray/protocol/outbound/VmessOutSettings.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/share/VmessShare.java b/server/src/main/java/io/wdd/func/xray/beans/xray/share/VmessShare.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HTTPRequestObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HTTPRequestObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HTTPResponseObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HTTPResponseObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HttpHeaderObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/HttpHeaderObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/InboundObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/InboundObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/MuxObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/MuxObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/NoneHeaderObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/NoneHeaderObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/OutboundObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/OutboundObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/StreamSettingsObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/StreamSettingsObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TcpObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TcpObject.java old mode 100644 new mode 100755 index 4abdcbd..b8d919c --- a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TcpObject.java +++ b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TcpObject.java @@ -1,6 +1,5 @@ package io.wdd.func.xray.beans.xray.transport; -import io.wdd.func.xray.beans.xray.transport.TCPHeaderObject; import lombok.Data; /** diff --git a/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TransportObject.java b/server/src/main/java/io/wdd/func/xray/beans/xray/transport/TransportObject.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/RealXrayPathConfig.json b/server/src/main/java/io/wdd/func/xray/persisit/RealXrayPathConfig.json old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/RealXrayPathConfig.txt b/server/src/main/java/io/wdd/func/xray/persisit/RealXrayPathConfig.txt old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/XrayDBOperator.java b/server/src/main/java/io/wdd/func/xray/persisit/XrayDBOperator.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/XrayOssOperator.java b/server/src/main/java/io/wdd/func/xray/persisit/XrayOssOperator.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/cache/EdgeWeightCache.java b/server/src/main/java/io/wdd/func/xray/persisit/cache/EdgeWeightCache.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/persisit/cache/ProxyNodeCache.java b/server/src/main/java/io/wdd/func/xray/persisit/cache/ProxyNodeCache.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/service/XrayCallAgent.java b/server/src/main/java/io/wdd/func/xray/service/XrayCallAgent.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/service/XrayConfigPersistor.java b/server/src/main/java/io/wdd/func/xray/service/XrayConfigPersistor.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/service/XrayCoreService.java b/server/src/main/java/io/wdd/func/xray/service/XrayCoreService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/func/xray/service/XrayCoreServiceImpl.java b/server/src/main/java/io/wdd/func/xray/service/XrayCoreServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/MessagePusher.java b/server/src/main/java/io/wdd/pusher/MessagePusher.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/beans/CmiiUpdateMessage.java b/server/src/main/java/io/wdd/pusher/beans/CmiiUpdateMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/beans/ImageSyncMessage.java b/server/src/main/java/io/wdd/pusher/beans/ImageSyncMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/beans/NtfyReceiveMessage.java b/server/src/main/java/io/wdd/pusher/beans/NtfyReceiveMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/beans/NtfySendMessage.java b/server/src/main/java/io/wdd/pusher/beans/NtfySendMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/pusher/beans/PushMessageType.java b/server/src/main/java/io/wdd/pusher/beans/PushMessageType.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/agent/AgentOperationMessage.java b/server/src/main/java/io/wdd/rpc/agent/AgentOperationMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/agent/AgentOperationType.java b/server/src/main/java/io/wdd/rpc/agent/AgentOperationType.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/agent/OctopusAgentService.java b/server/src/main/java/io/wdd/rpc/agent/OctopusAgentService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/agent/OctopusAgentServiceImpl.java b/server/src/main/java/io/wdd/rpc/agent/OctopusAgentServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/beans/request/ExecutionEntity.java b/server/src/main/java/io/wdd/rpc/beans/request/ExecutionEntity.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/beans/request/MetricQueryEntity.java b/server/src/main/java/io/wdd/rpc/beans/request/MetricQueryEntity.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/config/RedisConfig.java b/server/src/main/java/io/wdd/rpc/config/RedisConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/config/RedisConfiguration.java b/server/src/main/java/io/wdd/rpc/config/RedisConfiguration.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/controller/AgentController.java b/server/src/main/java/io/wdd/rpc/controller/AgentController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/controller/ExecutionController.java b/server/src/main/java/io/wdd/rpc/controller/ExecutionController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/controller/SchedulerController.java b/server/src/main/java/io/wdd/rpc/controller/SchedulerController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/controller/StatusController.java b/server/src/main/java/io/wdd/rpc/controller/StatusController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/execute/ExecutionMessage.java b/server/src/main/java/io/wdd/rpc/execute/ExecutionMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/execute/ExecutionMessageType.java b/server/src/main/java/io/wdd/rpc/execute/ExecutionMessageType.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/execute/service/ExecutionService.java b/server/src/main/java/io/wdd/rpc/execute/service/ExecutionService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/execute/service/ExecutionServiceImpl.java b/server/src/main/java/io/wdd/rpc/execute/service/ExecutionServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/init/AcceptAgentInitInfo.java b/server/src/main/java/io/wdd/rpc/init/AcceptAgentInitInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/init/InitRabbitMQConfig.java b/server/src/main/java/io/wdd/rpc/init/InitRabbitMQConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/init/OctopusExchangeConfig.java b/server/src/main/java/io/wdd/rpc/init/OctopusExchangeConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/OctopusMessage.java b/server/src/main/java/io/wdd/rpc/message/OctopusMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/OctopusMessageType.java b/server/src/main/java/io/wdd/rpc/message/OctopusMessageType.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/handler/OMessageHandler.java b/server/src/main/java/io/wdd/rpc/message/handler/OMessageHandler.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/handler/OMessageReplayContent.java b/server/src/main/java/io/wdd/rpc/message/handler/OMessageReplayContent.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/handler/OMessageToServerListener.java b/server/src/main/java/io/wdd/rpc/message/handler/OMessageToServerListener.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/message/sender/OMessageToAgentSender.java b/server/src/main/java/io/wdd/rpc/message/sender/OMessageToAgentSender.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/beans/OctopusQuartzJob.java b/server/src/main/java/io/wdd/rpc/scheduler/beans/OctopusQuartzJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/beans/OctopusQuartzLog.java b/server/src/main/java/io/wdd/rpc/scheduler/beans/OctopusQuartzLog.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/beans/ScriptSchedulerDTO.java b/server/src/main/java/io/wdd/rpc/scheduler/beans/ScriptSchedulerDTO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/beans/ScriptSchedulerVO.java b/server/src/main/java/io/wdd/rpc/scheduler/beans/ScriptSchedulerVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/config/ExecutionJob.java b/server/src/main/java/io/wdd/rpc/scheduler/config/ExecutionJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzLogOperator.java b/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzLogOperator.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzRunnable.java b/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzRunnable.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzSchedulerUtils.java b/server/src/main/java/io/wdd/rpc/scheduler/config/QuartzSchedulerUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/dto/ScheduleUpdateDBInfoJob.java b/server/src/main/java/io/wdd/rpc/scheduler/dto/ScheduleUpdateDBInfoJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/job/AgentAliveStatusMonitorJob.java b/server/src/main/java/io/wdd/rpc/scheduler/job/AgentAliveStatusMonitorJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/job/AgentMetricStatusJob.java b/server/src/main/java/io/wdd/rpc/scheduler/job/AgentMetricStatusJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/job/AgentScriptSchedulerJob.java b/server/src/main/java/io/wdd/rpc/scheduler/job/AgentScriptSchedulerJob.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/BuildStatusScheduleTask.java b/server/src/main/java/io/wdd/rpc/scheduler/service/BuildStatusScheduleTask.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/QuartzSchedulerService.java b/server/src/main/java/io/wdd/rpc/scheduler/service/QuartzSchedulerService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/QuartzSchedulerServiceImpl.java b/server/src/main/java/io/wdd/rpc/scheduler/service/QuartzSchedulerServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/script/AgentApplyScheduledScript.java b/server/src/main/java/io/wdd/rpc/scheduler/service/script/AgentApplyScheduledScript.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentAliveStatusMonitorService.java b/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentAliveStatusMonitorService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentMetricStatusCollectService.java b/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentMetricStatusCollectService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentStatusStreamReader.java b/server/src/main/java/io/wdd/rpc/scheduler/service/status/AgentStatusStreamReader.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/scheduler/定时任务说明.md b/server/src/main/java/io/wdd/rpc/scheduler/定时任务说明.md old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/AgentHealthyStatusEnum.java b/server/src/main/java/io/wdd/rpc/status/AgentHealthyStatusEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/AllAgentStatusCache.java b/server/src/main/java/io/wdd/rpc/status/AllAgentStatusCache.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/OctopusStatusMessage.java b/server/src/main/java/io/wdd/rpc/status/OctopusStatusMessage.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/AgentMetric.java b/server/src/main/java/io/wdd/rpc/status/beans/AgentMetric.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/CPUMetric.java b/server/src/main/java/io/wdd/rpc/status/beans/CPUMetric.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/DiskInfo.java b/server/src/main/java/io/wdd/rpc/status/beans/DiskInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/DockerMetric.java b/server/src/main/java/io/wdd/rpc/status/beans/DockerMetric.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/MemoryMetric.java b/server/src/main/java/io/wdd/rpc/status/beans/MemoryMetric.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/NetworkMetric.java b/server/src/main/java/io/wdd/rpc/status/beans/NetworkMetric.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/beans/OMessageStatusTypeEnum.java b/server/src/main/java/io/wdd/rpc/status/beans/OMessageStatusTypeEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/AgentStatus.java b/server/src/main/java/io/wdd/rpc/status/deprecate/AgentStatus.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/AgentSystemInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/AgentSystemInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/AppStatusEnum.java b/server/src/main/java/io/wdd/rpc/status/deprecate/AppStatusEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/AppStatusInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/AppStatusInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/CpuInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/CpuInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/CpuTicks.java b/server/src/main/java/io/wdd/rpc/status/deprecate/CpuTicks.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/DiskInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/DiskInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/MemoryInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/MemoryInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/MetricStatus.java b/server/src/main/java/io/wdd/rpc/status/deprecate/MetricStatus.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/deprecate/NetworkInfo.java b/server/src/main/java/io/wdd/rpc/status/deprecate/NetworkInfo.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/service/SyncStatusService.java b/server/src/main/java/io/wdd/rpc/status/service/SyncStatusService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/rpc/status/service/SyncStatusServiceImpl.java b/server/src/main/java/io/wdd/rpc/status/service/SyncStatusServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/mapper/ProjectInfoDTOMapper.java b/server/src/main/java/io/wdd/server/beans/mapper/ProjectInfoDTOMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/mapper/ServerInfoDTOMapper.java b/server/src/main/java/io/wdd/server/beans/mapper/ServerInfoDTOMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/mapper/ServerRoleDTOMapper.java b/server/src/main/java/io/wdd/server/beans/mapper/ServerRoleDTOMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/po/AppInfoPO.java b/server/src/main/java/io/wdd/server/beans/po/AppInfoPO.java old mode 100644 new mode 100755 index 264128a..9ce68d3 --- a/server/src/main/java/io/wdd/server/beans/po/AppInfoPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/AppInfoPO.java @@ -1,13 +1,14 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.*; +import com.baomidou.mybatisplus.annotation.FieldFill; +import com.baomidou.mybatisplus.annotation.TableField; +import com.baomidou.mybatisplus.annotation.TableId; +import com.baomidou.mybatisplus.annotation.TableName; +import com.fasterxml.jackson.annotation.JsonFormat; +import lombok.Data; import java.io.Serializable; import java.time.LocalDateTime; -import java.util.Date; - -import com.fasterxml.jackson.annotation.JsonFormat; -import lombok.Data; /** * diff --git a/server/src/main/java/io/wdd/server/beans/po/AppdomainInfoPO.java b/server/src/main/java/io/wdd/server/beans/po/AppdomainInfoPO.java old mode 100644 new mode 100755 index e71e463..40ab306 --- a/server/src/main/java/io/wdd/server/beans/po/AppdomainInfoPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/AppdomainInfoPO.java @@ -1,13 +1,14 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.*; +import com.baomidou.mybatisplus.annotation.FieldFill; +import com.baomidou.mybatisplus.annotation.TableField; +import com.baomidou.mybatisplus.annotation.TableId; +import com.baomidou.mybatisplus.annotation.TableName; +import com.fasterxml.jackson.annotation.JsonFormat; +import lombok.Data; import java.io.Serializable; import java.time.LocalDateTime; -import java.util.Date; - -import com.fasterxml.jackson.annotation.JsonFormat; -import lombok.Data; /** * diff --git a/server/src/main/java/io/wdd/server/beans/po/DomainInfoPO.java b/server/src/main/java/io/wdd/server/beans/po/DomainInfoPO.java old mode 100644 new mode 100755 index db81dd9..94bc60f --- a/server/src/main/java/io/wdd/server/beans/po/DomainInfoPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/DomainInfoPO.java @@ -1,13 +1,14 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.*; +import com.baomidou.mybatisplus.annotation.FieldFill; +import com.baomidou.mybatisplus.annotation.TableField; +import com.baomidou.mybatisplus.annotation.TableId; +import com.baomidou.mybatisplus.annotation.TableName; +import com.fasterxml.jackson.annotation.JsonFormat; +import lombok.Data; import java.io.Serializable; import java.time.LocalDateTime; -import java.util.Date; - -import com.fasterxml.jackson.annotation.JsonFormat; -import lombok.Data; /** * diff --git a/server/src/main/java/io/wdd/server/beans/po/ExecutionLogPO.java b/server/src/main/java/io/wdd/server/beans/po/ExecutionLogPO.java old mode 100644 new mode 100755 index 804e59e..83df081 --- a/server/src/main/java/io/wdd/server/beans/po/ExecutionLogPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/ExecutionLogPO.java @@ -1,13 +1,12 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.IdType; import com.baomidou.mybatisplus.annotation.TableField; import com.baomidou.mybatisplus.annotation.TableId; import com.baomidou.mybatisplus.annotation.TableName; +import lombok.Data; + import java.io.Serializable; import java.time.LocalDateTime; -import java.util.Date; -import lombok.Data; /** * diff --git a/server/src/main/java/io/wdd/server/beans/po/ProjectInfoPO.java b/server/src/main/java/io/wdd/server/beans/po/ProjectInfoPO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/po/ProjectServerRelationPO.java b/server/src/main/java/io/wdd/server/beans/po/ProjectServerRelationPO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/po/ScriptSchedulerPO.java b/server/src/main/java/io/wdd/server/beans/po/ScriptSchedulerPO.java old mode 100644 new mode 100755 index c067dc8..3424b97 --- a/server/src/main/java/io/wdd/server/beans/po/ScriptSchedulerPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/ScriptSchedulerPO.java @@ -1,18 +1,17 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.IdType; import com.baomidou.mybatisplus.annotation.TableField; import com.baomidou.mybatisplus.annotation.TableId; import com.baomidou.mybatisplus.annotation.TableName; -import java.io.Serializable; -import java.time.LocalDateTime; - import io.swagger.annotations.ApiModel; import lombok.AllArgsConstructor; import lombok.Data; import lombok.NoArgsConstructor; import lombok.experimental.SuperBuilder; +import java.io.Serializable; +import java.time.LocalDateTime; + /** * * @TableName script_scheduler diff --git a/server/src/main/java/io/wdd/server/beans/po/ServerAppRelationPO.java b/server/src/main/java/io/wdd/server/beans/po/ServerAppRelationPO.java old mode 100644 new mode 100755 index 20f6269..41cd412 --- a/server/src/main/java/io/wdd/server/beans/po/ServerAppRelationPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/ServerAppRelationPO.java @@ -1,12 +1,11 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.IdType; import com.baomidou.mybatisplus.annotation.TableField; -import com.baomidou.mybatisplus.annotation.TableId; import com.baomidou.mybatisplus.annotation.TableName; -import java.io.Serializable; import lombok.Data; +import java.io.Serializable; + /** * * @TableName server_app_relation diff --git a/server/src/main/java/io/wdd/server/beans/po/ServerDomainRelationPO.java b/server/src/main/java/io/wdd/server/beans/po/ServerDomainRelationPO.java old mode 100644 new mode 100755 index a2282ce..05e0468 --- a/server/src/main/java/io/wdd/server/beans/po/ServerDomainRelationPO.java +++ b/server/src/main/java/io/wdd/server/beans/po/ServerDomainRelationPO.java @@ -1,12 +1,11 @@ package io.wdd.server.beans.po; -import com.baomidou.mybatisplus.annotation.IdType; import com.baomidou.mybatisplus.annotation.TableField; -import com.baomidou.mybatisplus.annotation.TableId; import com.baomidou.mybatisplus.annotation.TableName; -import java.io.Serializable; import lombok.Data; +import java.io.Serializable; + /** * * @TableName server_domain_relation diff --git a/server/src/main/java/io/wdd/server/beans/po/ServerInfoPO.java b/server/src/main/java/io/wdd/server/beans/po/ServerInfoPO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/po/ServerInfoRoleEnum.java b/server/src/main/java/io/wdd/server/beans/po/ServerInfoRoleEnum.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/po/ServerRolePO.java b/server/src/main/java/io/wdd/server/beans/po/ServerRolePO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/request/ProjectQueryEntity.java b/server/src/main/java/io/wdd/server/beans/request/ProjectQueryEntity.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/request/ServerQueryEntity.java b/server/src/main/java/io/wdd/server/beans/request/ServerQueryEntity.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/vo/AppInfoVO.java b/server/src/main/java/io/wdd/server/beans/vo/AppInfoVO.java old mode 100644 new mode 100755 index 8a5ba9a..ff8c16f --- a/server/src/main/java/io/wdd/server/beans/vo/AppInfoVO.java +++ b/server/src/main/java/io/wdd/server/beans/vo/AppInfoVO.java @@ -1,14 +1,14 @@ package io.wdd.server.beans.vo; -import java.time.LocalDateTime; - import com.fasterxml.jackson.annotation.JsonFormat; import lombok.AllArgsConstructor; import lombok.Data; import lombok.NoArgsConstructor; import lombok.experimental.SuperBuilder; +import java.time.LocalDateTime; + /** * (AppInfo)实体类对应的VO类 * diff --git a/server/src/main/java/io/wdd/server/beans/vo/AppdomainInfoVO.java b/server/src/main/java/io/wdd/server/beans/vo/AppdomainInfoVO.java old mode 100644 new mode 100755 index 19890f7..802bfee --- a/server/src/main/java/io/wdd/server/beans/vo/AppdomainInfoVO.java +++ b/server/src/main/java/io/wdd/server/beans/vo/AppdomainInfoVO.java @@ -2,15 +2,12 @@ package io.wdd.server.beans.vo; import com.baomidou.mybatisplus.annotation.FieldFill; import com.baomidou.mybatisplus.annotation.TableField; -import com.baomidou.mybatisplus.annotation.TableId; -import com.baomidou.mybatisplus.annotation.TableName; import com.fasterxml.jackson.annotation.JsonFormat; import lombok.AllArgsConstructor; import lombok.Data; import lombok.NoArgsConstructor; import lombok.experimental.SuperBuilder; -import java.io.Serializable; import java.time.LocalDateTime; /** diff --git a/server/src/main/java/io/wdd/server/beans/vo/DomainInfoVO.java b/server/src/main/java/io/wdd/server/beans/vo/DomainInfoVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/vo/ProjectInfoVO.java b/server/src/main/java/io/wdd/server/beans/vo/ProjectInfoVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/vo/ProjectServerVO.java b/server/src/main/java/io/wdd/server/beans/vo/ProjectServerVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/vo/ServerInfoVO.java b/server/src/main/java/io/wdd/server/beans/vo/ServerInfoVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/beans/vo/ServerRoleVO.java b/server/src/main/java/io/wdd/server/beans/vo/ServerRoleVO.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/config/Knife4jConfig.java b/server/src/main/java/io/wdd/server/config/Knife4jConfig.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/config/ServerCommonPool.java b/server/src/main/java/io/wdd/server/config/ServerCommonPool.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/AppController.java b/server/src/main/java/io/wdd/server/controller/AppController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/DomainController.java b/server/src/main/java/io/wdd/server/controller/DomainController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/ProjectController.java b/server/src/main/java/io/wdd/server/controller/ProjectController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/ProjectServerController.java b/server/src/main/java/io/wdd/server/controller/ProjectServerController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/RoleController.java b/server/src/main/java/io/wdd/server/controller/RoleController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/controller/ServerController.java b/server/src/main/java/io/wdd/server/controller/ServerController.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreAppService.java b/server/src/main/java/io/wdd/server/coreService/CoreAppService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreDomainService.java b/server/src/main/java/io/wdd/server/coreService/CoreDomainService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreProjectServerService.java b/server/src/main/java/io/wdd/server/coreService/CoreProjectServerService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreProjectService.java b/server/src/main/java/io/wdd/server/coreService/CoreProjectService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreRoleService.java b/server/src/main/java/io/wdd/server/coreService/CoreRoleService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/CoreServerService.java b/server/src/main/java/io/wdd/server/coreService/CoreServerService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreAppServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreAppServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreDomainServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreDomainServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreProjectServerServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreProjectServerServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreProjectServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreProjectServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreRoleServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreRoleServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/coreService/impl/CoreServerServiceImpl.java b/server/src/main/java/io/wdd/server/coreService/impl/CoreServerServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/mapper/AppInfoMapper.java b/server/src/main/java/io/wdd/server/mapper/AppInfoMapper.java old mode 100644 new mode 100755 index 969110e..4b3f000 --- a/server/src/main/java/io/wdd/server/mapper/AppInfoMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/AppInfoMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.AppInfoPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.AppInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/AppdomainInfoMapper.java b/server/src/main/java/io/wdd/server/mapper/AppdomainInfoMapper.java old mode 100644 new mode 100755 index 4b5ed59..3adbc2d --- a/server/src/main/java/io/wdd/server/mapper/AppdomainInfoMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/AppdomainInfoMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.AppdomainInfoPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.AppdomainInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/DomainInfoMapper.java b/server/src/main/java/io/wdd/server/mapper/DomainInfoMapper.java old mode 100644 new mode 100755 index 6190cef..52c42e2 --- a/server/src/main/java/io/wdd/server/mapper/DomainInfoMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/DomainInfoMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.DomainInfoPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.DomainInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/ExecutionLogMapper.java b/server/src/main/java/io/wdd/server/mapper/ExecutionLogMapper.java old mode 100644 new mode 100755 index 7c596ec..eb520fc --- a/server/src/main/java/io/wdd/server/mapper/ExecutionLogMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/ExecutionLogMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.ExecutionLogPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.ExecutionLogPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/ProjectInfoMapper.java b/server/src/main/java/io/wdd/server/mapper/ProjectInfoMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/mapper/ProjectServerRelationMapper.java b/server/src/main/java/io/wdd/server/mapper/ProjectServerRelationMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/mapper/ScriptSchedulerMapper.java b/server/src/main/java/io/wdd/server/mapper/ScriptSchedulerMapper.java old mode 100644 new mode 100755 index dbf4949..e5d515e --- a/server/src/main/java/io/wdd/server/mapper/ScriptSchedulerMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/ScriptSchedulerMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.ScriptSchedulerPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.ScriptSchedulerPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/ServerAppRelationMapper.java b/server/src/main/java/io/wdd/server/mapper/ServerAppRelationMapper.java old mode 100644 new mode 100755 index 39c4a14..a631d92 --- a/server/src/main/java/io/wdd/server/mapper/ServerAppRelationMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/ServerAppRelationMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.ServerAppRelationPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.ServerAppRelationPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/ServerDomainRelationMapper.java b/server/src/main/java/io/wdd/server/mapper/ServerDomainRelationMapper.java old mode 100644 new mode 100755 index 3cd7866..d763ac6 --- a/server/src/main/java/io/wdd/server/mapper/ServerDomainRelationMapper.java +++ b/server/src/main/java/io/wdd/server/mapper/ServerDomainRelationMapper.java @@ -1,7 +1,7 @@ package io.wdd.server.mapper; -import io.wdd.server.beans.po.ServerDomainRelationPO; import com.baomidou.mybatisplus.core.mapper.BaseMapper; +import io.wdd.server.beans.po.ServerDomainRelationPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/mapper/ServerInfoMapper.java b/server/src/main/java/io/wdd/server/mapper/ServerInfoMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/mapper/ServerRoleMapper.java b/server/src/main/java/io/wdd/server/mapper/ServerRoleMapper.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/AppInfoService.java b/server/src/main/java/io/wdd/server/service/AppInfoService.java old mode 100644 new mode 100755 index de52d9a..20d9ac3 --- a/server/src/main/java/io/wdd/server/service/AppInfoService.java +++ b/server/src/main/java/io/wdd/server/service/AppInfoService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.AppInfoPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.AppInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/AppdomainInfoService.java b/server/src/main/java/io/wdd/server/service/AppdomainInfoService.java old mode 100644 new mode 100755 index 7b0076f..ffeb5cc --- a/server/src/main/java/io/wdd/server/service/AppdomainInfoService.java +++ b/server/src/main/java/io/wdd/server/service/AppdomainInfoService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.AppdomainInfoPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.AppdomainInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/DomainInfoService.java b/server/src/main/java/io/wdd/server/service/DomainInfoService.java old mode 100644 new mode 100755 index a75cb93..f6e9576 --- a/server/src/main/java/io/wdd/server/service/DomainInfoService.java +++ b/server/src/main/java/io/wdd/server/service/DomainInfoService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.DomainInfoPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.DomainInfoPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/ExecutionLogService.java b/server/src/main/java/io/wdd/server/service/ExecutionLogService.java old mode 100644 new mode 100755 index 594a729..3ec5a81 --- a/server/src/main/java/io/wdd/server/service/ExecutionLogService.java +++ b/server/src/main/java/io/wdd/server/service/ExecutionLogService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.ExecutionLogPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.ExecutionLogPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/ProjectInfoService.java b/server/src/main/java/io/wdd/server/service/ProjectInfoService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/ProjectServerRelationService.java b/server/src/main/java/io/wdd/server/service/ProjectServerRelationService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/ScriptSchedulerService.java b/server/src/main/java/io/wdd/server/service/ScriptSchedulerService.java old mode 100644 new mode 100755 index 041bbfa..f02346b --- a/server/src/main/java/io/wdd/server/service/ScriptSchedulerService.java +++ b/server/src/main/java/io/wdd/server/service/ScriptSchedulerService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.ScriptSchedulerPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.ScriptSchedulerPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/ServerAppRelationService.java b/server/src/main/java/io/wdd/server/service/ServerAppRelationService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/ServerDomainRelationService.java b/server/src/main/java/io/wdd/server/service/ServerDomainRelationService.java old mode 100644 new mode 100755 index ab807ea..5a8e0db --- a/server/src/main/java/io/wdd/server/service/ServerDomainRelationService.java +++ b/server/src/main/java/io/wdd/server/service/ServerDomainRelationService.java @@ -1,7 +1,7 @@ package io.wdd.server.service; -import io.wdd.server.beans.po.ServerDomainRelationPO; import com.baomidou.mybatisplus.extension.service.IService; +import io.wdd.server.beans.po.ServerDomainRelationPO; /** * @author wdd diff --git a/server/src/main/java/io/wdd/server/service/ServerInfoService.java b/server/src/main/java/io/wdd/server/service/ServerInfoService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/ServerRoleService.java b/server/src/main/java/io/wdd/server/service/ServerRoleService.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/impl/AppInfoServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/AppInfoServiceImpl.java old mode 100644 new mode 100755 index 8a49c7e..8b8fb1e --- a/server/src/main/java/io/wdd/server/service/impl/AppInfoServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/AppInfoServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.AppInfoPO; -import io.wdd.server.service.AppInfoService; import io.wdd.server.mapper.AppInfoMapper; +import io.wdd.server.service.AppInfoService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/AppdomainInfoServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/AppdomainInfoServiceImpl.java old mode 100644 new mode 100755 index 6d66279..bdea494 --- a/server/src/main/java/io/wdd/server/service/impl/AppdomainInfoServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/AppdomainInfoServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.AppdomainInfoPO; -import io.wdd.server.service.AppdomainInfoService; import io.wdd.server.mapper.AppdomainInfoMapper; +import io.wdd.server.service.AppdomainInfoService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/DomainInfoServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/DomainInfoServiceImpl.java old mode 100644 new mode 100755 index 0117790..b31e481 --- a/server/src/main/java/io/wdd/server/service/impl/DomainInfoServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/DomainInfoServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.DomainInfoPO; -import io.wdd.server.service.DomainInfoService; import io.wdd.server.mapper.DomainInfoMapper; +import io.wdd.server.service.DomainInfoService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/ExecutionLogServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ExecutionLogServiceImpl.java old mode 100644 new mode 100755 index 4c173a5..ef60f2f --- a/server/src/main/java/io/wdd/server/service/impl/ExecutionLogServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/ExecutionLogServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.ExecutionLogPO; -import io.wdd.server.service.ExecutionLogService; import io.wdd.server.mapper.ExecutionLogMapper; +import io.wdd.server.service.ExecutionLogService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/ProjectInfoServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ProjectInfoServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/impl/ProjectServerRelationServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ProjectServerRelationServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/impl/ScriptSchedulerServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ScriptSchedulerServiceImpl.java old mode 100644 new mode 100755 index e6c8074..15b4dbe --- a/server/src/main/java/io/wdd/server/service/impl/ScriptSchedulerServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/ScriptSchedulerServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.ScriptSchedulerPO; -import io.wdd.server.service.ScriptSchedulerService; import io.wdd.server.mapper.ScriptSchedulerMapper; +import io.wdd.server.service.ScriptSchedulerService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/ServerAppRelationServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ServerAppRelationServiceImpl.java old mode 100644 new mode 100755 index 0071c44..cbfa27b --- a/server/src/main/java/io/wdd/server/service/impl/ServerAppRelationServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/ServerAppRelationServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.ServerAppRelationPO; -import io.wdd.server.service.ServerAppRelationService; import io.wdd.server.mapper.ServerAppRelationMapper; +import io.wdd.server.service.ServerAppRelationService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/ServerDomainRelationServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ServerDomainRelationServiceImpl.java old mode 100644 new mode 100755 index d3cc86e..eca0fde --- a/server/src/main/java/io/wdd/server/service/impl/ServerDomainRelationServiceImpl.java +++ b/server/src/main/java/io/wdd/server/service/impl/ServerDomainRelationServiceImpl.java @@ -2,8 +2,8 @@ package io.wdd.server.service.impl; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import io.wdd.server.beans.po.ServerDomainRelationPO; -import io.wdd.server.service.ServerDomainRelationService; import io.wdd.server.mapper.ServerDomainRelationMapper; +import io.wdd.server.service.ServerDomainRelationService; import org.springframework.stereotype.Service; /** diff --git a/server/src/main/java/io/wdd/server/service/impl/ServerInfoServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ServerInfoServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/service/impl/ServerRoleServiceImpl.java b/server/src/main/java/io/wdd/server/service/impl/ServerRoleServiceImpl.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/utils/DaemonDatabaseOperator.java b/server/src/main/java/io/wdd/server/utils/DaemonDatabaseOperator.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/utils/EntityConvertUtils.java b/server/src/main/java/io/wdd/server/utils/EntityConvertUtils.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/utils/MyBatisAutoInsertInterceptor.java b/server/src/main/java/io/wdd/server/utils/MyBatisAutoInsertInterceptor.java old mode 100644 new mode 100755 diff --git a/server/src/main/java/io/wdd/server/utils/SpringUtils.java b/server/src/main/java/io/wdd/server/utils/SpringUtils.java old mode 100644 new mode 100755 index 4a0f514..927bd18 --- a/server/src/main/java/io/wdd/server/utils/SpringUtils.java +++ b/server/src/main/java/io/wdd/server/utils/SpringUtils.java @@ -7,7 +7,6 @@ import org.springframework.beans.factory.config.BeanFactoryPostProcessor; import org.springframework.beans.factory.config.ConfigurableListableBeanFactory; import org.springframework.context.ApplicationContext; import org.springframework.context.ApplicationContextAware; -import org.springframework.data.redis.stream.StreamMessageListenerContainer; import org.springframework.stereotype.Component; diff --git a/server/src/main/java/io/wdd/server/核心功能设计.md b/server/src/main/java/io/wdd/server/核心功能设计.md old mode 100644 new mode 100755 diff --git a/server/src/main/resources/META-INF/additional-spring-configuration-metadata.json b/server/src/main/resources/META-INF/additional-spring-configuration-metadata.json old mode 100644 new mode 100755 diff --git a/server/src/main/resources/application-local.yml b/server/src/main/resources/application-local.yml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/application.yml b/server/src/main/resources/application.yml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/bootstrap.yml-back b/server/src/main/resources/bootstrap.yml-back old mode 100644 new mode 100755 diff --git a/server/src/main/resources/customMapper/ProjectServerRelationMapperCustom.xml b/server/src/main/resources/customMapper/ProjectServerRelationMapperCustom.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/AgentStatusCollector.xml b/server/src/main/resources/mapper/AgentStatusCollector.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/AppInfoMapper.xml b/server/src/main/resources/mapper/AppInfoMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/AppdomainInfoMapper.xml b/server/src/main/resources/mapper/AppdomainInfoMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ExecutionLogMapper.xml b/server/src/main/resources/mapper/ExecutionLogMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ProjectInfoMapper.xml b/server/src/main/resources/mapper/ProjectInfoMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ProjectServerRelationMapper.xml b/server/src/main/resources/mapper/ProjectServerRelationMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/QuartzSchedulerService.xml b/server/src/main/resources/mapper/QuartzSchedulerService.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ScriptSchedulerMapper.xml b/server/src/main/resources/mapper/ScriptSchedulerMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ServerDomainRelationMapper.xml b/server/src/main/resources/mapper/ServerDomainRelationMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ServerInfoMapper.xml b/server/src/main/resources/mapper/ServerInfoMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/mapper/ServerRoleMapper.xml b/server/src/main/resources/mapper/ServerRoleMapper.xml old mode 100644 new mode 100755 diff --git a/server/src/main/resources/xrayResult/1.json b/server/src/main/resources/xrayResult/1.json old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/ServerApplicationTests.java b/server/src/test/java/io/wdd/server/ServerApplicationTests.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/SimpleTest.java b/server/src/test/java/io/wdd/server/SimpleTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/controller/ProjectInfoTest.java b/server/src/test/java/io/wdd/server/controller/ProjectInfoTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/controller/ProjectServerTest.java b/server/src/test/java/io/wdd/server/controller/ProjectServerTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/controller/ServerInfoTest.java b/server/src/test/java/io/wdd/server/controller/ServerInfoTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/controller/ServerRoleTest.java b/server/src/test/java/io/wdd/server/controller/ServerRoleTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/coreService/TestCoreProjectService.java b/server/src/test/java/io/wdd/server/coreService/TestCoreProjectService.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/func/ExecutionEntityJson.json b/server/src/test/java/io/wdd/server/func/ExecutionEntityJson.json old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/func/TestBaseFuncScheduler.java b/server/src/test/java/io/wdd/server/func/TestBaseFuncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/func/TestImageSyncScheduler.java b/server/src/test/java/io/wdd/server/func/TestImageSyncScheduler.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/func/TestSingleCommand.java b/server/src/test/java/io/wdd/server/func/TestSingleCommand.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/func/XrayGenerateTest.java b/server/src/test/java/io/wdd/server/func/XrayGenerateTest.java old mode 100644 new mode 100755 diff --git a/server/src/test/java/io/wdd/server/pusher/PusherTest.java b/server/src/test/java/io/wdd/server/pusher/PusherTest.java old mode 100644 new mode 100755 diff --git a/setting-aliyun.xml b/setting-aliyun.xml old mode 100644 new mode 100755 diff --git a/socks5_txthinking/.gitignore b/socks5_txthinking/.gitignore old mode 100644 new mode 100755 diff --git a/socks5_txthinking/LICENSE b/socks5_txthinking/LICENSE old mode 100644 new mode 100755 diff --git a/socks5_txthinking/README.md b/socks5_txthinking/README.md old mode 100644 new mode 100755 diff --git a/socks5_txthinking/README_ZH.md b/socks5_txthinking/README_ZH.md old mode 100644 new mode 100755 diff --git a/socks5_txthinking/bind.go b/socks5_txthinking/bind.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/build_all.ps1 b/socks5_txthinking/build_all.ps1 old mode 100644 new mode 100755 diff --git a/socks5_txthinking/client.go b/socks5_txthinking/client.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/client_side.go b/socks5_txthinking/client_side.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/connect.go b/socks5_txthinking/connect.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/example_test.go b/socks5_txthinking/example_test.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/go.mod b/socks5_txthinking/go.mod old mode 100644 new mode 100755 diff --git a/socks5_txthinking/go.sum b/socks5_txthinking/go.sum old mode 100644 new mode 100755 diff --git a/socks5_txthinking/init.go b/socks5_txthinking/init.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/main.go b/socks5_txthinking/main.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/old_tcp_tailscale/logger.go b/socks5_txthinking/old_tcp_tailscale/logger.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/old_tcp_tailscale/main.go b/socks5_txthinking/old_tcp_tailscale/main.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/old_tcp_tailscale/socks5.go b/socks5_txthinking/old_tcp_tailscale/socks5.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/runnergrounp.go b/socks5_txthinking/runnergrounp.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/server.go b/socks5_txthinking/server.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/server_side.go b/socks5_txthinking/server_side.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/socks5.go b/socks5_txthinking/socks5.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/udp.go b/socks5_txthinking/udp.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/util.go b/socks5_txthinking/util.go old mode 100644 new mode 100755 diff --git a/socks5_txthinking/util_test.go b/socks5_txthinking/util_test.go old mode 100644 new mode 100755 diff --git a/source/.gitignore b/source/.gitignore old mode 100644 new mode 100755 diff --git a/source/pom.xml b/source/pom.xml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/Dockerfile b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/Dockerfile old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/docker-compose.yml b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/docker-compose.yml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/.bitnami_components.json b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/.bitnami_components.json old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/licenses/licenses.txt b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/licenses/licenses.txt old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libbitnami.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libbitnami.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libfile.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libfile.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libfs.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libfs.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libhook.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libhook.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/liblog.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/liblog.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libnet.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libnet.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libos.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libos.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libpersistence.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libpersistence.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libservice.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libservice.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libvalidations.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libvalidations.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libversion.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libversion.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libwebserver.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/opt/bitnami/scripts/libwebserver.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/usr/sbin/install_packages b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/prebuildfs/usr/sbin/install_packages old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/libredis.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/libredis.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis-env.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis-env.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/entrypoint.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/entrypoint.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/postunpack.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/postunpack.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/run.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/run.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/setup.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/rootfs/opt/bitnami/scripts/redis/setup.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/tags-info.yaml b/source/src/main/java/io/wdd/source/bitnamiredisshell/bitnami-redis-6.2/tags-info.yaml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/entrypoint.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/entrypoint.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libbitnami.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libbitnami.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libfile.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libfile.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libfs.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libfs.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libhook.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libhook.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/liblog.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/liblog.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libnet.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libnet.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libos.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libos.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libpersistence.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libpersistence.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libredis.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libredis.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libservice.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libservice.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libvalidations.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libvalidations.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libversion.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libversion.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libwebserver.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/libwebserver.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/postunpack.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/postunpack.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/redis-env.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/redis-env.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/run.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/run.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/setup.sh b/source/src/main/java/io/wdd/source/bitnamiredisshell/manual/setup.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/build-my-own-base-image.dockerfile b/source/src/main/java/io/wdd/source/build-my-own-base-image.dockerfile old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus-middile-wares.zip b/source/src/main/java/io/wdd/source/octopus-middile-wares.zip old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/0-cleanup-k3s.sh b/source/src/main/java/io/wdd/source/octopus/0-cleanup-k3s.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/nacos-2.2.0-initdb-schema.sql b/source/src/main/java/io/wdd/source/octopus/nacos-2.2.0-initdb-schema.sql old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/nginx-reverse.sh b/source/src/main/java/io/wdd/source/octopus/nginx-reverse.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/openai/openai-web-dockercompose.yaml b/source/src/main/java/io/wdd/source/octopus/openai/openai-web-dockercompose.yaml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/openai/openai-wmm-nginx.conf b/source/src/main/java/io/wdd/source/octopus/openai/openai-wmm-nginx.conf old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/openai/说明.txt b/source/src/main/java/io/wdd/source/octopus/openai/说明.txt old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/run-enviroment-compose.yaml b/source/src/main/java/io/wdd/source/octopus/run-enviroment-compose.yaml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/octopus/simple-middleware.yaml b/source/src/main/java/io/wdd/source/octopus/simple-middleware.yaml old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/agent-bootup.sh b/source/src/main/java/io/wdd/source/shell/agent-bootup.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/agent-restart.sh b/source/src/main/java/io/wdd/source/shell/agent-restart.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/agent-shutdown.sh b/source/src/main/java/io/wdd/source/shell/agent-shutdown.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/agent-update.sh b/source/src/main/java/io/wdd/source/shell/agent-update.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/env/hosts.conf b/source/src/main/java/io/wdd/source/shell/env/hosts.conf old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/env/resolv.conf b/source/src/main/java/io/wdd/source/shell/env/resolv.conf old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/function/execute-scrpit-loadCPUtoAllMachine.json b/source/src/main/java/io/wdd/source/shell/function/execute-scrpit-loadCPUtoAllMachine.json old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/function/load50cpu_2_hour.sh b/source/src/main/java/io/wdd/source/shell/function/load50cpu_2_hour.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/install-nginx.sh b/source/src/main/java/io/wdd/source/shell/install-nginx.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-env.sh b/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-env.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-file.sh b/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-file.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-log.sh b/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-log.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-sys.sh b/source/src/main/java/io/wdd/source/shell/lib/wdd-lib-sys.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/readme.md b/source/src/main/java/io/wdd/source/shell/readme.md old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/server-bootup.sh b/source/src/main/java/io/wdd/source/shell/server-bootup.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/source/shell/test.sh b/source/src/main/java/io/wdd/source/shell/test.sh old mode 100644 new mode 100755 diff --git a/source/src/main/java/io/wdd/wdd_server.sql b/source/src/main/java/io/wdd/wdd_server.sql old mode 100644 new mode 100755