ArangoDB v3.8 reached End of Life (EOL) and is no longer supported.

This documentation is outdated. Please see the most recent version at docs.arangodb.com

ArangoDeployment Custom Resource

The ArangoDB Deployment Operator creates and maintains ArangoDB deployments in a Kubernetes cluster, given a deployment specification. This deployment specification is a CustomResource following a CustomResourceDefinition created by the operator.

Example minimal deployment definition of an ArangoDB database cluster:

apiVersion: "database.arangodb.com/v1"
kind: "ArangoDeployment"
metadata:
  name: "example-arangodb-cluster"
spec:
  mode: Cluster

Example more elaborate deployment definition:

apiVersion: "database.arangodb.com/v1"
kind: "ArangoDeployment"
metadata:
  name: "example-arangodb-cluster"
spec:
  mode: Cluster
  environment: Production
  agents:
    count: 3
    args:
      - --log.level=debug
    resources:
      requests:
        storage: 8Gi
    storageClassName: ssd
  dbservers:
    count: 5
    resources:
      requests:
        storage: 80Gi
    storageClassName: ssd
  coordinators:
    count: 3
  image: "arangodb/arangodb:3.9.3"

Specification reference

Below you’ll find all settings of the ArangoDeployment custom resource. Several settings are for various groups of servers. These are indicated with <group> where <group> can be any of:

  • agents for all Agents of a Cluster or ActiveFailover pair.
  • dbservers for all DB-Servers of a Cluster.
  • coordinators for all Coordinators of a Cluster.
  • single for all single servers of a Single instance or ActiveFailover pair.
  • syncmasters for all syncmasters of a Cluster.
  • syncworkers for all syncworkers of a Cluster.

Special group id can be used for image discovery and testing affinity/toleration settings.

spec.architecture: []string

This setting specifies a CPU architecture for the deployment. Possible values are:

  • amd64 (default): Use processors with the x86-64 architecture.
  • arm64: Use processors with the 64-bit ARM architecture.

The setting expects a list of strings, but you should only specify a single list item for the architecture, except when you want to migrate from one architecture to the other. The first list item defines the new default architecture for the deployment that you want to migrate to.

To use the ARM architecture, you need to enable it in the operator first using --set "operator.architectures={amd64,arm64}". See Installation with Helm.

To create a new deployment with arm64 nodes, specify the architecture in the deployment specification as follows:

spec:
  architecture:
    - arm64

To migrate nodes of an existing deployment from amd64 to arm64, modify the deployment specification so that both architectures are listed:

 spec:
   architecture:
+    - arm64
     - amd64

This lets new members as well as recreated members use arm64 nodes.

Then run the following command:

kubectl annotate pod $POD "deployment.arangodb.com/replace=true"

To change an existing member to arm64, annotate the pod as follows:

kubectl annotate pod $POD "deployment.arangodb.com/arch=arm64"

An ArchitectureMismatch condition occurs in the deployment:

members:
  single:
    - arango-version: 3.10.0
      architecture: arm64
      conditions:
        reason: Member has a different architecture than the deployment
        status: "True"
        type: ArchitectureMismatch

Restart the pod using this command:

kubectl annotate pod $POD "deployment.arangodb.com/rotate=true"

spec.mode: string

This setting specifies the type of deployment you want to create. Possible values are:

  • Cluster (default) Full cluster. Defaults to 3 Agents, 3 DB-Servers & 3 Coordinators.
  • ActiveFailover Active-failover single pair. Defaults to 3 Agents and 2 single servers.
  • Single Single server only (note this does not provide high availability or reliability).

This setting cannot be changed after the deployment has been created.

spec.environment: string

This setting specifies the type of environment in which the deployment is created. Possible values are:

  • Development (default) This value optimizes the deployment for development use. It is possible to run a deployment on a small number of nodes (e.g. minikube).
  • Production This value optimizes the deployment for production use. It puts required affinity constraints on all pods to avoid Agents & DB-Servers from running on the same machine.

spec.image: string

This setting specifies the docker image to use for all ArangoDB servers. In a development environment this setting defaults to arangodb/arangodb:latest. For production environments this is a required setting without a default value. It is highly recommend to use explicit version (not latest) for production environments.

spec.imagePullPolicy: string

This setting specifies the pull policy for the docker image to use for all ArangoDB servers. Possible values are:

  • IfNotPresent (default) to pull only when the image is not found on the node.
  • Always to always pull the image before using it.

spec.imagePullSecrets: []string

This setting specifies the list of image pull secrets for the docker image to use for all ArangoDB servers.

spec.annotations: map[string]string

This setting set specified annotations to all ArangoDeployment owned resources (pods, services, PVC’s, PDB’s).

spec.storageEngine: string

This setting specifies the type of storage engine used for all servers in the cluster. Possible values are:

  • MMFiles To use the MMFiles storage engine.
  • RocksDB (default) To use the RocksDB storage engine.

This setting cannot be changed after the cluster has been created.

spec.downtimeAllowed: bool

This setting is used to allow automatic reconciliation actions that yield some downtime of the ArangoDB deployment. When this setting is set to false (the default), no automatic action that may result in downtime is allowed. If the need for such an action is detected, an event is added to the ArangoDeployment.

Once this setting is set to true, the automatic action is executed.

Operations that may result in downtime are:

  • Rotating TLS CA certificate

Note: It is still possible that there is some downtime when the Kubernetes cluster is down, or in a bad state, irrespective of the value of this setting.

spec.memberPropagationMode

Changes to a pod’s configuration require a restart of that pod in almost all cases. Pods are restarted eagerly by default, which can cause more restarts than desired, especially when updating arangod as well as the operator. The propagation of the configuration changes can be deferred to the next restart, either triggered manually by the user or by another operation like an upgrade. This reduces the number of restarts for upgrading both the server and the operator from two to one.

  • always: Restart the member as soon as a configuration change is discovered
  • on-restart: Wait until the next restart to change the member configuration

spec.rocksdb.encryption.keySecretName

This setting specifies the name of a Kubernetes Secret that contains an encryption key used for encrypting all data stored by ArangoDB servers. When an encryption key is used, encryption of the data in the cluster is enabled, without it encryption is disabled. The default value is empty.

This requires the Enterprise Edition.

The encryption key cannot be changed after the cluster has been created.

The secret specified by this setting, must have a data field named ‘key’ containing an encryption key that is exactly 32 bytes long.

spec.networkAttachedVolumes: bool

The default of this option is false. If set to true, a ResignLeaderShip operation will be triggered when a DB-Server pod is evicted (rather than a CleanOutServer operation). Furthermore, the pod will simply be redeployed on a different node, rather than cleaned and retired and replaced by a new member. You must only set this option to true if your persistent volumes are “movable” in the sense that they can be mounted from a different k8s node, like in the case of network attached volumes. If your persistent volumes are tied to a specific pod, you must leave this option on false.

spec.externalAccess.type: string

This setting specifies the type of Service that will be created to provide access to the ArangoDB deployment from outside the Kubernetes cluster. Possible values are:

  • None To limit access to application running inside the Kubernetes cluster.
  • LoadBalancer To create a Service of type LoadBalancer for the ArangoDB deployment.
  • NodePort To create a Service of type NodePort for the ArangoDB deployment.
  • Auto (default) To create a Service of type LoadBalancer and fallback to a Service or type NodePort when the LoadBalancer is not assigned an IP address.

spec.externalAccess.loadBalancerIP: string

This setting specifies the IP used to for the LoadBalancer to expose the ArangoDB deployment on. This setting is used when spec.externalAccess.type is set to LoadBalancer or Auto.

If you do not specify this setting, an IP will be chosen automatically by the load-balancer provisioner.

spec.externalAccess.loadBalancerSourceRanges: []string

If specified and supported by the platform (cloud provider), this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature.

More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/

spec.externalAccess.nodePort: int

This setting specifies the port used to expose the ArangoDB deployment on. This setting is used when spec.externalAccess.type is set to NodePort or Auto.

If you do not specify this setting, a random port will be chosen automatically.

spec.externalAccess.advertisedEndpoint: string

This setting specifies the advertised endpoint for all Coordinators.

spec.auth.jwtSecretName: string

This setting specifies the name of a kubernetes Secret that contains the JWT token used for accessing all ArangoDB servers. When no name is specified, it defaults to <deployment-name>-jwt. To disable authentication, set this value to None.

If you specify a name of a Secret, that secret must have the token in a data field named token.

If you specify a name of a Secret that does not exist, a random token is created and stored in a Secret with given name.

Changing a JWT token results in stopping the entire cluster and restarting it.

spec.tls.caSecretName: string

This setting specifies the name of a kubernetes Secret that contains a standard CA certificate + private key used to sign certificates for individual ArangoDB servers. When no name is specified, it defaults to <deployment-name>-ca. To disable authentication, set this value to None.

If you specify a name of a Secret that does not exist, a self-signed CA certificate + key is created and stored in a Secret with given name.

The specified Secret, must contain the following data fields:

  • ca.crt PEM encoded public key of the CA certificate
  • ca.key PEM encoded private key of the CA certificate

spec.tls.altNames: []string

This setting specifies a list of alternate names that will be added to all generated certificates. These names can be DNS names or email addresses. The default value is empty.

spec.tls.ttl: duration

This setting specifies the time to live of all generated server certificates. The default value is 2160h (about 3 month).

When the server certificate is about to expire, it will be automatically replaced by a new one and the affected server will be restarted.

Note: The time to live of the CA certificate (when created automatically) will be set to 10 years.

spec.sync.enabled: bool

This setting enables/disables support for data center 2 data center replication in the cluster. When enabled, the cluster will contain a number of syncmaster & syncworker servers. The default value is false.

spec.sync.externalAccess.type: string

This setting specifies the type of Service that will be created to provide access to the ArangoSync syncMasters from outside the Kubernetes cluster. Possible values are:

  • None To limit access to applications running inside the Kubernetes cluster.
  • LoadBalancer To create a Service of type LoadBalancer for the ArangoSync SyncMasters.
  • NodePort To create a Service of type NodePort for the ArangoSync SyncMasters.
  • Auto (default) To create a Service of type LoadBalancer and fallback to a Service or type NodePort when the LoadBalancer is not assigned an IP address.

Note that when you specify a value of None, a Service will still be created, but of type ClusterIP.

spec.sync.externalAccess.loadBalancerIP: string

This setting specifies the IP used for the LoadBalancer to expose the ArangoSync SyncMasters on. This setting is used when spec.sync.externalAccess.type is set to LoadBalancer or Auto.

If you do not specify this setting, an IP will be chosen automatically by the load-balancer provisioner.

spec.sync.externalAccess.nodePort: int

This setting specifies the port used to expose the ArangoSync SyncMasters on. This setting is used when spec.sync.externalAccess.type is set to NodePort or Auto.

If you do not specify this setting, a random port will be chosen automatically.

spec.sync.externalAccess.loadBalancerSourceRanges: []string

If specified and supported by the platform (cloud provider), this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature.

More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/

spec.sync.externalAccess.masterEndpoint: []string

This setting specifies the master endpoint(s) advertised by the ArangoSync SyncMasters. If not set, this setting defaults to:

  • If spec.sync.externalAccess.loadBalancerIP is set, it defaults to https://<load-balancer-ip>:<8629>.
  • Otherwise it defaults to https://<sync-service-dns-name>:<8629>.

spec.sync.externalAccess.accessPackageSecretNames: []string

This setting specifies the names of zero of more Secrets that will be created by the deployment operator containing “access packages”. An access package contains those Secrets that are needed to access the SyncMasters of this ArangoDeployment.

By removing a name from this setting, the corresponding Secret is also deleted. Note that to remove all access packages, leave an empty array in place ([]). Completely removing the setting results in not modifying the list.

See the ArangoDeploymentReplication specification for more information on access packages.

spec.sync.auth.jwtSecretName: string

This setting specifies the name of a kubernetes Secret that contains the JWT token used for accessing all ArangoSync master servers. When not specified, the spec.auth.jwtSecretName value is used.

If you specify a name of a Secret that does not exist, a random token is created and stored in a Secret with given name.

spec.sync.auth.clientCASecretName: string

This setting specifies the name of a kubernetes Secret that contains a PEM encoded CA certificate used for client certificate verification in all ArangoSync master servers. This is a required setting when spec.sync.enabled is true. The default value is empty.

spec.sync.mq.type: string

This setting sets the type of message queue used by ArangoSync. Possible values are:

  • Direct (default) for direct HTTP connections between the 2 data centers.

spec.sync.tls.caSecretName: string

This setting specifies the name of a kubernetes Secret that contains a standard CA certificate + private key used to sign certificates for individual ArangoSync master servers.

When no name is specified, it defaults to <deployment-name>-sync-ca.

If you specify a name of a Secret that does not exist, a self-signed CA certificate + key is created and stored in a Secret with given name.

The specified Secret, must contain the following data fields:

  • ca.crt PEM encoded public key of the CA certificate
  • ca.key PEM encoded private key of the CA certificate

spec.sync.tls.altNames: []string

This setting specifies a list of alternate names that will be added to all generated certificates. These names can be DNS names or email addresses. The default value is empty.

spec.sync.monitoring.tokenSecretName: string

This setting specifies the name of a kubernetes Secret that contains the bearer token used for accessing all monitoring endpoints of all ArangoSync servers. When not specified, no monitoring token is used. The default value is empty.

spec.disableIPv6: bool

This setting prevents the use of IPv6 addresses by ArangoDB servers. The default is false.

This setting cannot be changed after the deployment has been created.

spec.restoreFrom: string

This setting specifies a ArangoBackup resource name the cluster should be restored from.

After a restore or failure to do so, the status of the deployment contains information about the restore operation in the restore key.

It will contain some of the following fields:

  • requestedFrom: name of the ArangoBackup used to restore from.
  • message: optional message explaining why the restore failed.
  • state: state indicating if the restore was successful or not. Possible values: Restoring, Restored, RestoreFailed

If the restoreFrom key is removed from the spec, the restore key is deleted as well.

A new restore attempt is made if and only if either in the status restore is not set or if spec.restoreFrom and status.requestedFrom are different.

spec.license.secretName: string

This setting specifies the name of a kubernetes Secret that contains the license key token used for enterprise images. This value is not used for the Community Edition.

spec.bootstrap.passwordSecretNames.root: string

This setting specifies a secret name for the credentials of the root user.

When a deployment is created the operator will setup the root user account according to the credentials given by the secret. If the secret doesn’t exist the operator creates a secret with a random password.

There are two magic values for the secret name:

  • None specifies no action. This disables root password randomization. This is the default value. (Thus the root password is empty - not recommended)
  • Auto specifies automatic name generation, which is <deploymentname>-root-password.

spec.metrics.enabled: bool

If this is set to true, the operator runs a sidecar container for every Agent, DB-Server, Coordinator and Single server.

In addition to the sidecar containers the operator will deploy a service to access the exporter ports (from within the k8s cluster), and a resource of type ServiceMonitor, provided the corresponding custom resource definition is deployed in the k8s cluster. If you are running Prometheus in the same k8s cluster with the Prometheus operator, this will be the case. The ServiceMonitor will have the following labels set:

  • app: arangodb
  • arango_deployment: YOUR_DEPLOYMENT_NAME
  • context: metrics
  • metrics: prometheus

This makes it possible that you configure your Prometheus deployment to automatically start monitoring on the available Prometheus feeds. To this end, you must configure the serviceMonitorSelector in the specs of your Prometheus deployment to match these labels. For example:

  serviceMonitorSelector:
    matchLabels:
      metrics: prometheus

would automatically select all pods of all ArangoDB cluster deployments which have metrics enabled.

spec.metrics.image: string

Deprecated in: v1.2.0 (kube-arangodb)

See above, this is the name of the Docker image for the ArangoDB exporter to expose metrics. If empty, the same image as for the main deployment is used.

spec.metrics.resources: ResourceRequirements

Introduced in: v0.4.3 (kube-arangodb)

This setting specifies the resources required by the metrics container. This includes requests and limits. See Kubernetes documentation.

spec.metrics.mode: string

Introduced in: v1.0.2 (kube-arangodb)

Defines metrics exporter mode.

Possible values:

  • exporter (default): add sidecar to pods (except Agency pods) and exposes metrics collected by exporter from ArangoDB Container. Exporter in this mode exposes metrics which are accessible without authentication.
  • sidecar: add sidecar to all pods and expose metrics from ArangoDB metrics endpoint. Exporter in this mode exposes metrics which are accessible without authentication.
  • internal: configure ServiceMonitor to use internal ArangoDB metrics endpoint (proper JWT token is generated for this endpoint).

spec.metrics.tls: bool

Introduced in: v1.1.0 (kube-arangodb)

Defines if TLS should be enabled on Metrics exporter endpoint. The default is true.

This option will enable TLS only if TLS is enabled on ArangoDeployment, otherwise true value will not take any effect.

spec.lifecycle.resources: ResourceRequirements

Introduced in: v0.4.3 (kube-arangodb)

This setting specifies the resources required by the lifecycle init container. This includes requests and limits. See Kubernetes documentation.

spec.<group>.count: number

This setting specifies the number of servers to start for the given group. For the Agent group, this value must be a positive, odd number. The default value is 3 for all groups except single (there the default is 1 for spec.mode: Single and 2 for spec.mode: ActiveFailover).

For the syncworkers group, it is highly recommended to use the same number as for the dbservers group.

spec.<group>.minCount: number

Specifies a minimum for the count of servers. If set, a specification is invalid if count < minCount.

spec.<group>.maxCount: number

Specifies a maximum for the count of servers. If set, a specification is invalid if count > maxCount.

spec.<group>.args: []string

This setting specifies additional command-line arguments passed to all servers of this group. The default value is an empty array.

spec.<group>.resources: ResourceRequirements

This setting specifies the resources required by pods of this group. This includes requests and limits.

See https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ for details.

spec.<group>.overrideDetectedTotalMemory: bool

Introduced in: v1.0.1 (kube-arangodb)

Set additional flag in ArangoDeployment pods to propagate Memory resource limits

spec.<group>.volumeClaimTemplate.Spec: PersistentVolumeClaimSpec

Specifies a volumeClaimTemplate used by operator to create to volume claims for pods of this group. This setting is not available for group coordinators, syncmasters & syncworkers.

The default value describes a volume with 8Gi storage, ReadWriteOnce access mode and volume mode set to PersistentVolumeFilesystem.

If this field is not set and spec.<group>.resources.requests.storage is set, then a default volume claim with size as specified by spec.<group>.resources.requests.storage will be created. In that case storage and iops is not forwarded to the pods resource requirements.

spec.<group>.pvcResizeMode: string

Specifies a resize mode used by operator to resize PVCs and PVs.

Supported modes:

  • runtime (default) - PVC will be resized in Pod runtime (EKS, GKE)
  • rotate - Pod will be shutdown and PVC will be resized (AKS)

spec.<group>.serviceAccountName: string

This setting specifies the serviceAccountName for the Pods created for each server of this group. If empty, it defaults to using the default service account.

Using an alternative ServiceAccount is typically used to separate access rights. The ArangoDB deployments need some very minimal access rights. With the deployment of the operator, we grant the following rights for the default service account:

rules:
  - apiGroups:
      - ""
    resources:
      - pods
    verbs:
      - get

If you are using a different service account, please grant these rights to that service account.

spec.<group>.annotations: map[string]string

This setting set annotations overrides for pods in this group. Annotations are merged with spec.annotations.

spec.<group>.priorityClassName: string

Priority class name for pods of this group. Will be forwarded to the pod spec. Kubernetes documentation

spec.<group>.probes.livenessProbeDisabled: bool

If set to true, the operator does not generate a liveness probe for new pods belonging to this group.

spec.<group>.probes.livenessProbeSpec.initialDelaySeconds: int

Number of seconds after the container has started before liveness or readiness probes are initiated. Defaults to 2 seconds. Minimum value is 0.

spec.<group>.probes.livenessProbeSpec.periodSeconds: int

How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.

spec.<group>.probes.livenessProbeSpec.timeoutSeconds: int

Number of seconds after which the probe times out. Defaults to 2 second. Minimum value is 1.

spec.<group>.probes.livenessProbeSpec.failureThreshold: int

When a Pod starts and the probe fails, Kubernetes will try failureThreshold times before giving up. Giving up means restarting the container. Defaults to 3. Minimum value is 1.

spec.<group>.probes.readinessProbeDisabled: bool

If set to true, the operator does not generate a readiness probe for new pods belonging to this group.

spec.<group>.probes.readinessProbeSpec.initialDelaySeconds: int

Number of seconds after the container has started before liveness or readiness probes are initiated. Defaults to 2 seconds. Minimum value is 0.

spec.<group>.probes.readinessProbeSpec.periodSeconds: int

How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.

spec.<group>.probes.readinessProbeSpec.timeoutSeconds: int

Number of seconds after which the probe times out. Defaults to 2 second. Minimum value is 1.

spec.<group>.probes.readinessProbeSpec.successThreshold: int

Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Minimum value is 1.

spec.<group>.probes.readinessProbeSpec.failureThreshold: int

When a Pod starts and the probe fails, Kubernetes will try failureThreshold times before giving up. Giving up means the Pod will be marked Unready. Defaults to 3. Minimum value is 1.

spec.<group>.allowMemberRecreation: bool

Introduced in: v1.2.1 (kube-arangodb)

This setting changes the member recreation logic based on group:

  • For Sync Masters, Sync Workers, Coordinator and DB-Servers it determines if a member can be recreated in case of failure (default true)
  • For Agents and Single this value is hardcoded to false and the value provided in spec is ignored.

spec.<group>.tolerations: []Toleration

This setting specifies the tolerations for the Pods created for each server of this group.

By default, suitable tolerations are set for the following keys with the NoExecute effect:

  • node.kubernetes.io/not-ready
  • node.kubernetes.io/unreachable
  • node.alpha.kubernetes.io/unreachable (will be removed in future version)

For more information on tolerations, consult the Kubernetes documentation.

spec.<group>.nodeSelector: map[string]string

This setting specifies a set of labels to be used as nodeSelector for Pods of this node.

For more information on node selectors, consult the Kubernetes documentation.

spec.<group>.entrypoint: string

Entrypoint overrides container executable.

spec.<group>.antiAffinity: PodAntiAffinity

Specifies additional antiAffinity settings in ArangoDB Pod definitions.

For more information on antiAffinity, consult the Kubernetes documentation.

spec.<group>.affinity: PodAffinity

Specifies additional affinity settings in ArangoDB Pod definitions.

For more information on affinity, consult the Kubernetes documentation.

spec.<group>.nodeAffinity: NodeAffinity

Specifies additional nodeAffinity settings in ArangoDB Pod definitions.

For more information on nodeAffinity, consult the Kubernetes documentation.

spec.<group>.securityContext: ServerGroupSpecSecurityContext

Specifies additional securityContext settings in ArangoDB Pod definitions. This is similar (but not fully compatible) to k8s SecurityContext definition.

For more information on securityContext, consult the Kubernetes documentation.

spec.<group>.securityContext.addCapabilities: []Capability

Adds new capabilities to containers.

spec.<group>.securityContext.allowPrivilegeEscalation: bool

Controls whether a process can gain more privileges than its parent process.

spec.<group>.securityContext.privileged: bool

Runs container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host.

spec.<group>.securityContext.readOnlyRootFilesystem: bool

Mounts the container’s root filesystem as read-only.

spec.<group>.securityContext.runAsNonRoot: bool

Indicates that the container must run as a non-root user.

spec.<group>.securityContext.runAsUser: integer

The UID to run the entrypoint of the container process.

spec.<group>.securityContext.runAsGroup: integer

The GID to run the entrypoint of the container process.

spec.<group>.securityContext.supplementalGroups: []integer

A list of groups applied to the first process run in each container, in addition to the container’s primary GID, the fsGroup (if specified), and group memberships defined in the container image for the UID of the container process.

spec.<group>.securityContext.fsGroup: integer

A special supplemental group that applies to all containers in a pod.

spec.<group>.securityContext.seccompProfile: SeccompProfile

The seccomp options to use by the containers in this pod.

spec.<group>.securityContext.seLinuxOptions: SELinuxOptions

The SELinux context to be applied to all containers.

Image discovery group spec.id fields

Image discovery (id) group only supports the next subset of fields. Refer to according field documentation in spec.<group> description.

  • spec.id.entrypoint: string
  • spec.id.tolerations: []Toleration
  • spec.id.nodeSelector: map[string]string
  • spec.id.priorityClassName: string
  • spec.id.antiAffinity: PodAntiAffinity
  • spec.id.affinity: PodAffinity
  • spec.id.nodeAffinity: NodeAffinity
  • spec.id.serviceAccountName: string
  • spec.id.securityContext: ServerGroupSpecSecurityContext
  • spec.id.resources: ResourceRequirements

Deprecated Fields

spec.<group>.resources.requests.storage: storageUnit

This setting specifies the amount of storage required for each server of this group. The default value is 8Gi.

This setting is not available for group coordinators, syncmasters & syncworkers because servers in these groups do not need persistent storage.

Please use VolumeClaimTemplate from now on. This field is not considered if VolumeClaimTemplate is set. Note however, that the information in requests is completely handed over to the pod in this case.

spec.<group>.storageClassName: string

This setting specifies the storageClass for the PersistentVolumes created for each server of this group.

This setting is not available for group coordinators, syncmasters & syncworkers because servers in these groups do not need persistent storage.

Please use VolumeClaimTemplate from now on. This field is not considered if VolumeClaimTemplate is set. Note however, that the information in requests is completely handed over to the pod in this case.