@@ -9,7 +9,7 @@ CloudBees CI (formerly known as _CloudBees Core_) consists of three Docker image
For each image, all files other than UBI and native packages
are included in a `files.tar` marked with a SHA-256 checksum.
A version of CloudBees CI is given in the format `2.249.1.2`
A version of CloudBees CI is given in the format `2.249.2.3`
where the first three components are aligned with a Jenkins LTS.
The Helm chart is coversioned with `core-oc`.
The `core-mm` image typically shares the same version,
...
...
@@ -22,13 +22,13 @@ plus whatever other customizations are desired:
```yaml
OperationsCenter:
Image:
dockerImage:your-registry/core-oc:2.249.1.2
dockerImage:your-registry/core-oc:2.249.2.3
Master:
Image:
dockerImage:your-registry/core-mm:2.249.1.2
dockerImage:your-registry/core-mm:2.249.2.3
Agents:
Image:
dockerImage:your-registry/agent:2.249.1.2
dockerImage:your-registry/agent:2.249.2.3
```
and [install via Helm 3](https://docs.cloudbees.com/docs/cloudbees-core/latest/kubernetes-install-guide/installing-kubernetes-using-helm) using the local copy of the chart:
The command deploys CloudBees Core on the Kubernetes cluster in the default configuration. The [configuration](#configuration) section lists the parameters that can be configured during installation.
### Ingress Controller Installation
### Ingress Controller Installation
The chart is designed, so it can install an nginx-ingress controller.
The `nginx.ingress.Enabled` field controls ingress controller installation and setup.
The chart is designed, so it can install an nginx-ingress controller.
The `nginx.ingress.Enabled` field controls ingress controller installation and setup.
To install the chart with the release name `cloudbees-core` and hostname cloudbees-core.example.com.
```console
...
...
@@ -56,19 +66,18 @@ To uninstall/delete the `cloudbees-core` deployment:
$helm delete cloudbees-core
```
> **NOTE**: The current version of the CloudBees Core Helm Chart only manages the Operation Center.
Users should manage Managed Master using Operation Center.
Users should manage Managed Master using Operation Center.
The `helm delete` command stops the CloudBees Core deployment than removes the OperationsCenter Center.
The release is still stored in the Helm database, but it will now have the status deleted.
The `helm delete` command stops the CloudBees Core deployment than removes the OperationsCenter Center.
The release is still stored in the Helm database, but it will now have the status deleted.
If you wish to completely remove the release, use the following variation of the `helm delete` command.
```console
$helm delete cloudbees-core --purge
```
```
> **IMPORTANT**: The `helm delete` command does NOT remove the persistent volume claims as precaution against data losss.
You will need to use the `kubectl delete pvc` command to delete the persistent volumn claims.
You will need to use the `kubectl delete pvc` command to delete the persistent volumn claims.
The command removes all the Kubernetes components associated with the chart and deletes the release.
...
...
@@ -79,12 +88,12 @@ The easiest way to consult it is through the command `helm inspect values cloudb
Each property can override a default value with a value that specific to your Kubernetes cluster
You can provide this values using the `--set` flag on the Helm command line.
Helm also support merging values files together, so that you can create a YAML file for each environment.
Helm also support merging values files together, so that you can create a YAML file for each environment.
### Environment Property Value Files
Helm provides the option to use a custom property values file to override the default values set in the `values.yaml` file.
CloudBees recommends creating a custom properties file to override the default for your environments, instead of directly editing the included values.yaml file.
CloudBees recommends creating a custom properties file to override the default for your environments, instead of directly editing the included values.yaml file.
To use an environment property value file with Helm, use the -f option as shown in the following example:
@@ -93,3 +102,80 @@ You can download the latest version of the `example-values.yaml` file from Cloud
## Additional Documentation
CloudBees provides complete and more detail installation and operation documentation on the CloudBees web site at https://go.cloudbees.com/docs/cloudbees-core/cloud-install-guide/kubernetes-helm-install/
| Agents.Image.dockerImage | string | `"063356183961.dkr.ecr.us-east-1.amazonaws.com/ubi/unified-distribution/agent:2.249.2.3"` | Used to override the default docker image used for agents |
| Agents.SeparateNamespace.Create | bool | `false` | If true, the second namespace will be created when installing this chart. Otherwise, the existing namespace should be labeled with `cloudbees.com/role: agents` in order for network policies to work. |
| Agents.SeparateNamespace.Enabled | bool | `false` | If enabled, agents resources will be created in a separate namespace as well as bindings allowing masters to schedule them. |
| Agents.SeparateNamespace.Name | string | `nil` | Namespace where to create agents resources. Defaults to `${namespace}-builds` where `${namespace}` is the namespace where the chart is installed. |
| Hibernation.Enabled | bool | `false` | Whether to enable the [Hibernation](https://docs.cloudbees.com/docs/cloudbees-ci/latest/cloud-admin-guide/managing-masters#_hibernation_of_managed_masters) feature |
| Hibernation.Image.dockerImage | string | `"cloudbees/managed-master-hibernation-monitor:230.ee066a318539"` | Used to override the default docker image |
| Hibernation.Image.dockerPullPolicy | string | `"IfNotPresent"` | Used to override the default pull policy |
| Master.Enabled | bool | `true` | Whether to create the resources required to schedule masters. |
| Master.Image.dockerImage | string | `"063356183961.dkr.ecr.us-east-1.amazonaws.com/ubi/unified-distribution/core-mm:2.249.2.3"` | Used to override the default docker image |
| Master.JavaOpts | string | `nil` | Additional Java options to pass to managed masters. For example, setting up a JMX port |
| Master.OperationsCenterNamespace | string | `nil` | When deploying Master resources, this grants an Operations Center deployed in another namespace the right to deploy masters |
| NetworkPolicy.Enabled | bool | `false` | Enable only if the cluster supports it. Read the [documentation](https://kubernetes.io/docs/concepts/services-networking/network-policies/) to understand what this is about. |
| NetworkPolicy.JMXSelectors | list | `[]` | Custom selectors for accessing JMX port |
| NetworkPolicy.ingressControllerSelector | list | `[]` | Custom selector for the ingress-controller |
| OperationsCenter.AgentListenerPort | int | `50000` | Container port for agent listener traffic |
| OperationsCenter.Annotations | object | `{}` | Additional annotations to put on the pod running Operations Center |
| OperationsCenter.CSRF.ProxyCompatibility | bool | `false` | Proxy compatibility for the default CSRF issuer |
| OperationsCenter.ContainerPort | int | `8080` | Container port for http traffic |
| OperationsCenter.ContextPath | string | `"/cjoc"` | the path under which Operations Center will be accessible in the given host. |
| OperationsCenter.Enabled | bool | `true` | Disable for particular use case like setting up namespaces to host masters only |
| OperationsCenter.HealthProbeLivenessFailureThreshold | int | `12` | Threshold for liveness failure |
| OperationsCenter.HostName | string | `nil` | The hostname used to access Operations Center through the ingress controller. |
| OperationsCenter.Image.dockerImage | string | `"063356183961.dkr.ecr.us-east-1.amazonaws.com/ubi/unified-distribution/core-oc:2.249.2.3"` | Container image to use for Operations Center |
| OperationsCenter.ImagePullSecrets | string | `nil` | The name of the image pull secret to pull private docker images |
| OperationsCenter.Ingress.Annotations | object | `{"kubernetes.io/tls-acme":"false"}` | annotations to put on Ingress object |
| OperationsCenter.Ingress.Class | string | `"nginx"` | Ingress class to use for OC and MM ingresses Should be set to the same value as nginx-ingress.controller.ingressClass if enabled |
| OperationsCenter.Ingress.tls.Enable | bool | `false` | Set this to true in order to enable TLS on the ingress record |
| OperationsCenter.Ingress.tls.SecretName | string | `nil` | The name of the secret containing the certificate and private key to terminate TLS for the ingress |
| OperationsCenter.JavaOpts | string | `nil` | Additional java options to pass to the Operations Center |
| OperationsCenter.LoadBalancerIP | string | `nil` | Optionally assign a known public LB IP |
| OperationsCenter.LoadBalancerSourceRanges | list | `["0.0.0.0/0"]` | Only applicable when using `ServiceType: LoadBalancer` |
| OperationsCenter.NodeSelector | object | `{}` | Node labels and tolerations for pod assignment ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#nodeselector |
| OperationsCenter.Platform | string | `"standard"` | Enables specific settings depending on the platform platform specific values are: `eks`, `aws`, `gke`, `aks`, `pks`, `openshift`, `openshift4` Note: `openshift` maps to OpenShift 3.x |
| OperationsCenter.Protocol | string | `"http"` | the protocol used to access CJOC. Possible values are http/https. |
| OperationsCenter.Resources.Limits.Cpu | int | `1` | CPU limit to run Operations Center https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#meaning-of-cpu |
| OperationsCenter.Resources.Limits.Memory | string | `"2G"` | Memory limit to run Operations Center https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#meaning-of-memory |
| OperationsCenter.Resources.Requests.Cpu | int | `1` | CPU request to run Operations Center https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#meaning-of-cpu |
| OperationsCenter.Resources.Requests.Memory | string | `"2G"` | Memory request to run Operations Center https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/#meaning-of-memory |
| OperationsCenter.Route.tls.Enable | bool | `false` | Set this to true in OpenShift to terminate TLS at route level |
| OperationsCenter.ServiceAgentListenerPort | int | `50000` | Controls the service port where Operations Center TCP port for agents is exposed. Don't change this parameter unless you know what you are doing |
| OperationsCenter.ServiceAnnotations | object | `{}` | Additional annotations to put on the Operations Center service |
| OperationsCenter.ServicePort | int | `80` | Controls the service port where Operations Center http port is exposed. Don't change this parameter unless you know what you are doing |
| OperationsCenter.ServiceType | string | `"ClusterIP"` | Service Type. Defaults to ClusterIP, since we recommend using an ingress controller. |
| OperationsCenter.Tolerations | list | `[]` | Specify tolerations for the Operations Center pod. See [documentation](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) |
| Persistence.AccessMode | string | `"ReadWriteOnce"` | Access mode for the PVC ([doc](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#access-modes)) |
| Persistence.Annotations | object | `{}` | Annotations to put on the PVC |
| Persistence.Size | string | `"20Gi"` | Size of the Operations Center volume |
| Persistence.StorageClass | string | `nil` | Persistent Volume Storage Class for Jenkins Home If defined, storageClassName: <storageClass>. If set to "-", storageClassName: "", which disables dynamic provisioning. If undefined (the default) or set to null, the default storage class will be used, unless specified otherwise below. If setting OperationsCenter.Platform == gke, a storage class backed with SSD drives will be created by this chart and used automatically. |
| Persistence.mounts | list | `[]` | Additional volume mounts to attach to Operations Center container |
| Persistence.volumes | list | `[]` | Additional volumes to attach to Operations Center pod |
| PodSecurityPolicy.Annotations | object | `{}` | Additional annotations to put on the PodSecurityPolicy, e.g. AppArmor/Seccomp settings |
| PodSecurityPolicy.Enabled | bool | `false` | Enables [Pod Security Policies](https://kubernetes.io/docs/concepts/policy/pod-security-policy/) support Enable only if the cluster supports it. |
| ingress-nginx.Enabled | bool | `false` | Installs the [ingress-nginx](https://github.com/kubernetes/ingress-nginx/tree/master/charts/ingress-nginx) controller (optional). Enable this section if you don't have an existing installation of ingress-nginx controller Note: use `beta.kubernetes.io/os` when deploying on Kubernetes versions below 1.16 |
| nginx-ingress.Enabled | bool | `false` | Installs the [nginx-ingress](https://github.com/helm/charts/tree/master/stable/nginx-ingress) controller (optional). DEPRECATED - Use ingress-nginx section instead. Enable this section if you don't have an existing installation of nginx-ingress controller Note: use `beta.kubernetes.io/os` when deploying on Kubernetes versions below 1.16 |
| rbac.hibernationMonitorServiceAccountName | string | `"managed-master-hibernation-monitor"` | Name of the service account the Hibernation monitor will run as (if enabled) |
| rbac.install | bool | `true` | Install `role`/`rolebindings`/`serviceAccount`. If false (and rbac is enabled in the cluster anyway), provide valid names for `serviceAccountName`, `masterServiceAccountName` and `hibernationMonitorServiceAccountName` |
| rbac.masterServiceAccountName | string | `"jenkins"` | Name of the service account Jenkins masters will run as |
| rbac.serviceAccountName | string | `"cjoc"` | Name of the service account Operations Center will run as |
| sidecarinjector.Enabled | bool | `false` | Whether to enable installation of Sidecar Injector |
This chart bootstraps a CloudBees Core deployment on a [Kubernetes](http://kubernetes.io) cluster using the [Helm](https://helm.sh) package manager.
## Prerequisites
- Kubernetes 1.10 or higher
- Helm 2.12 or higher
{{ template "chart.requirementsSection" . }}
## Installing the Chart
### Default installation
To install the chart with the release name `cloudbees-core` and hostname `cloudbees-core.example.com`. The default installation requires nginx-ingress controller to be installed. The chart can install the nginx-ingress controller for you. This installation is described in the next section.
The command deploys CloudBees Core on the Kubernetes cluster in the default configuration. The [configuration](#configuration) section lists the parameters that can be configured during installation.
### Ingress Controller Installation
The chart is designed, so it can install an nginx-ingress controller.
The `nginx.ingress.Enabled` field controls ingress controller installation and setup.
To install the chart with the release name `cloudbees-core` and hostname cloudbees-core.example.com.
```console
$ helm install cloudbees/cloudbees-core \
--name cloudbees-core \
--set nginx.ingress.Enabled=true
```
## Uninstalling the Chart
To uninstall/delete the `cloudbees-core` deployment:
```console
$ helm delete cloudbees-core
```
> **NOTE**: The current version of the CloudBees Core Helm Chart only manages the Operation Center.
Users should manage Managed Master using Operation Center.
The `helm delete` command stops the CloudBees Core deployment than removes the OperationsCenter Center.
The release is still stored in the Helm database, but it will now have the status deleted.
If you wish to completely remove the release, use the following variation of the `helm delete` command.
```console
$ helm delete cloudbees-core --purge
```
> **IMPORTANT**: The `helm delete` command does NOT remove the persistent volume claims as precaution against data losss.
You will need to use the `kubectl delete pvc` command to delete the persistent volumn claims.
The command removes all the Kubernetes components associated with the chart and deletes the release.
## Configuration
Please refer to the chart `values.yaml` to get the exhaustive list of values that can be customized.
The easiest way to consult it is through the command `helm inspect values cloudbees/cloudbees-core`.
Each property can override a default value with a value that specific to your Kubernetes cluster
You can provide this values using the `--set` flag on the Helm command line.
Helm also support merging values files together, so that you can create a YAML file for each environment.
### Environment Property Value Files
Helm provides the option to use a custom property values file to override the default values set in the `values.yaml` file.
CloudBees recommends creating a custom properties file to override the default for your environments, instead of directly editing the included values.yaml file.
To use an environment property value file with Helm, use the -f option as shown in the following example:
You can download the latest version of the `example-values.yaml` file from CloudBees Examples GitHub repository at https://github.com/cloudbees/cloudbees-examples/tree/master/helm-custom-value-file-examples.
## Additional Documentation
CloudBees provides complete and more detail installation and operation documentation on the CloudBees web site at https://go.cloudbees.com/docs/cloudbees-core/cloud-install-guide/kubernetes-helm-install/
# Controls the service port where Operations Center http port is exposed.
# Don't change this parameter unless you know what you are doing
# OperationsCenter.ServicePort -- Controls the service port where Operations Center http port is exposed. Don't change this parameter unless you know what you are doing
ServicePort:80
# Controls the service port where Operations Center TCP port for agents is exposed.
# Don't change this parameter unless you know what you are doing
# OperationsCenter.ServiceAgentListenerPort -- Controls the service port where Operations Center TCP port for agents is exposed. Don't change this parameter unless you know what you are doing
ServiceAgentListenerPort:50000
# Service Type
# Defaults to ClusterIP, since we recommend using an ingress controller.
# OperationsCenter.ServiceType -- Service Type. Defaults to ClusterIP, since we recommend using an ingress controller.
ServiceType:ClusterIP
# Additional annotations to put on the Operations Center service
# OperationsCenter.ServiceAnnotations -- Additional annotations to put on the Operations Center service
# OperationsCenter.Tolerations -- Specify tolerations for the Operations Center pod.
# See [documentation](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/)
Tolerations:[]
# Additional annotations to put on the pod running Operations Center
# OperationsCenter.Annotations -- Additional annotations to put on the pod running Operations Center
Annotations:{}
Ingress:
# Ingress class to use for OC and MM ingresses
# OperationsCenter.Ingress.Class -- Ingress class to use for OC and MM ingresses
# Should be set to the same value as nginx-ingress.controller.ingressClass if enabled
Class:nginx
# Annotations to put on Ingress object
# OperationsCenter.Ingress.Annotations -- annotations to put on Ingress object
Annotations:
kubernetes.io/tls-acme:'false'
tls:
## Set this to true in order to enable TLS on the ingress record
# OperationsCenter.Ingress.tls.Enable -- Set this to true in order to enable TLS on the ingress record
Enable:false
SecretName:cloudbees-core-example-com-tls
Host:jenkins.cluster.local
# OperationsCenter.Ingress.tls.SecretName -- The name of the secret containing the certificate
# and private key to terminate TLS for the ingress
SecretName:null
# OpenShift uses a Route instead of an Ingress, this section is added to match the opensource Jenkins Chart
Route:
tls:
# OperationsCenter.Route.tls.Enable -- Set this to true in OpenShift to terminate TLS at route level
Enable:false
## @param ExtraConfigMaps - array of objects - optional
...
...
@@ -209,88 +233,91 @@ OperationsCenter:
# Master options
Master:
# Master.Enabled -- Whether to create the resources required to schedule masters.
Enabled:true
# When deploying Master resources, this grants an Operations Center deployed in another namespace the right to deploy masters
# Master.OperationsCenterNamespace -- When deploying Master resources, this grants an Operations Center deployed in another namespace the right to deploy masters
OperationsCenterNamespace:null
# Docker image inserted in Operations Center automatically
Image:
dockerImage:dcar/core-mm:2.249.1.2
dockerPullPolicy:IfNotPresent
# Additional Java options to pass to managed masters. For example, setting up a JMX port
# Master.Image.dockerImage -- Used to override the default docker image
dockerImage:dcar/core-mm:2.249.2.3
# Master.JavaOpts -- Additional Java options to pass to managed masters. For example, setting up a JMX port
# If enabled, agents resources will be created in a separate namespace as well as bindings allowing masters to schedule them.
# Agents.SeparateNamespace.Enabled -- If enabled, agents resources will be created in a separate namespace as well as bindings allowing masters to schedule them.
Enabled:false
# Namespace where to create agents resources. Defaults to ${namespace}-builds where ${namespace} is the namespace where the chart is installed.
# Agents.SeparateNamespace.Name -- Namespace where to create agents resources. Defaults to `${namespace}-builds` where `${namespace}` is the namespace where the chart is installed.
Name:null
# If true, the second namespace will be created when installing this chart.
# Otherwise, the existing namespace should be labeled with "cloudbees.com/role: agents" in order for network policies to work.
# Agents.SeparateNamespace.Create -- If true, the second namespace will be created when installing this chart. Otherwise, the existing namespace should be labeled with `cloudbees.com/role: agents` in order for network policies to work.
Create:false
Image:
# Used to override the default docker image used for agents
dockerImage:dcar/agent:2.249.1.2
# Agents.Image.dockerImage -- Used to override the default docker image used for agents
dockerImage:dcar/agent:2.249.2.3
Persistence:
## Persistent Volume Storage Class for Jenkins Home
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
##
## If undefined (the default) or set to null, the default storage class will be used,
## unless specified otherwise below.
##
## If setting OperationsCenter.Platform == gke, a storage class backed with
## SSD drives will be created by this chart and used automatically.
##
# Persistence.StorageClass -- Persistent Volume Storage Class for Jenkins Home
# If defined, storageClassName: <storageClass>.
# If set to "-", storageClassName: "", which disables dynamic provisioning.
# If undefined (the default) or set to null, the default storage class will be used,
# unless specified otherwise below.
# If setting OperationsCenter.Platform == gke, a storage class backed with
# SSD drives will be created by this chart and used automatically.
#
# StorageClass: "-"
StorageClass:null
# On AKS, use the following
# StorageClass: "managed-premium"
# Annotations to put on the PVC
# Persistence.Annotations -- Annotations to put on the PVC
Annotations:{}
# Persistence.AccessMode -- Access mode for the PVC ([doc](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#access-modes))
AccessMode:ReadWriteOnce
# Size of the Operations Center volume
# Persistence.Size -- Size of the Operations Center volume
Size:20Gi
# Additional volumes to attach to Operations Center pod
volumes:
# Persistence.volumes -- Additional volumes to attach to Operations Center pod
volumes:[]
# - name: nothing
# emptyDir: {}
# Additional volume mounts to attach to Operations Center container
mounts:
# Persistence.mounts -- Additional volume mounts to attach to Operations Center container
mounts:[]
# - mountPath: /var/nothing
# name: nothing
# readOnly: true
# Master hibernation
Hibernation:
# Hibernation.Enabled -- Whether to enable the [Hibernation](https://docs.cloudbees.com/docs/cloudbees-ci/latest/cloud-admin-guide/managing-masters#_hibernation_of_managed_masters) feature
Enabled:false
Image:
# Hibernation.Image.dockerImage -- Used to override the default docker image
# NetworkPolicy.Enabled -- Enable only if the cluster supports it. Read the [documentation](https://kubernetes.io/docs/concepts/services-networking/network-policies/) to understand what this is about.
Enabled:false
# Custom selector for the ingress-controller
ingressControllerSelector:
# NetworkPolicy.ingressControllerSelector -- Custom selector for the ingress-controller
ingressControllerSelector:[]
# - namespaceSelector:
# matchLabels:
# name: nginx-ingress
...
...
@@ -299,8 +326,8 @@ NetworkPolicy:
# app: nginx-ingress
# component: controller
# Custom selectors for accessing JMX port
JMXSelectors:
# NetworkPolicy.JMXSelectors -- Custom selectors for accessing JMX port
JMXSelectors:[]
# - namespaceSelector:
# matchLabels:
# name: datadog
...
...
@@ -310,23 +337,21 @@ NetworkPolicy:
## Install Default RBAC roles and bindings
rbac:
# Install role/rolebindings/serviceAccount.
# If false (and rbac is enabled in the cluster anyway), provide valid names for
# If false (and rbac is enabled in the cluster anyway), provide valid names for `serviceAccountName`, `masterServiceAccountName` and `hibernationMonitorServiceAccountName`