From 06315b3e2de826d983ab7200228f7b753e66620f Mon Sep 17 00:00:00 2001 From: jenkins-x-bot Date: Wed, 7 Feb 2024 11:18:07 +0000 Subject: [PATCH] chore: release 1.15.0 --- changelog.md | 6 + charts/lighthouse/Chart.yaml | 4 +- charts/lighthouse/README.md | 308 ++++++++++++------------ charts/lighthouse/lighthouse-1.15.0.tgz | Bin 0 -> 15774 bytes charts/lighthouse/values.yaml | 151 +----------- 5 files changed, 166 insertions(+), 303 deletions(-) create mode 100644 changelog.md create mode 100644 charts/lighthouse/lighthouse-1.15.0.tgz diff --git a/changelog.md b/changelog.md new file mode 100644 index 000000000..bc0bf50cb --- /dev/null +++ b/changelog.md @@ -0,0 +1,6 @@ + +## Changes in version 1.15.0 + +### New Features + +* allow to skip draft prs unless /ok-to-test is added (JordanGoasdoue) diff --git a/charts/lighthouse/Chart.yaml b/charts/lighthouse/Chart.yaml index edc18364e..c9ea016b8 100644 --- a/charts/lighthouse/Chart.yaml +++ b/charts/lighthouse/Chart.yaml @@ -3,6 +3,6 @@ description: | This chart bootstraps installation of [Lighthouse](https://github.com/jenkins-x/lighthouse). icon: https://raw.githubusercontent.com/jenkins-x/jenkins-x-website/master/images/logo/jenkinsx-icon-color.svg name: lighthouse -version: 0.1.0-SNAPSHOT +version: 1.15.0 home: https://github.com/jenkins-x/lighthouse - +appVersion: 1.15.0 diff --git a/charts/lighthouse/README.md b/charts/lighthouse/README.md index ed4700eef..48370b1c7 100644 --- a/charts/lighthouse/README.md +++ b/charts/lighthouse/README.md @@ -42,157 +42,161 @@ helm uninstall my-lighthouse --namespace lighthouse ## Values -| Key | Type | Description | Default | -| --------------------------------------------------- | ------ | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------- | -| `cluster.crds.create` | bool | Create custom resource definitions | `true` | -| `configMaps.config` | string | Raw `config.yaml` content | `nil` | -| `configMaps.configUpdater` | object | Settings used to configure the `config-updater` plugin | `{"orgAndRepo":"","path":""}` | -| `configMaps.create` | bool | Enables creation of `config.yaml` and `plugins.yaml` config maps | `false` | -| `configMaps.plugins` | string | Raw `plugins.yaml` content | `nil` | -| `engines.jenkins` | bool | Enables the Jenkins engine | `false` | -| `engines.jx` | bool | Enables the jx engine | `true` | -| `engines.tekton` | bool | Enables the tekton engine | `false` | -| `env` | object | Environment variables | `{"JX_DEFAULT_IMAGE":""}` | -| `externalPlugins[0].name` | string | | `"cd-indicators"` | -| `externalPlugins[0].requiredResources[0].kind` | string | | `"Service"` | -| `externalPlugins[0].requiredResources[0].name` | string | | `"cd-indicators"` | -| `externalPlugins[0].requiredResources[0].namespace` | string | | `"jx"` | -| `externalPlugins[1].name` | string | | `"lighthouse-webui-plugin"` | -| `externalPlugins[1].requiredResources[0].kind` | string | | `"Service"` | -| `externalPlugins[1].requiredResources[0].name` | string | | `"lighthouse-webui-plugin"` | -| `externalPlugins[1].requiredResources[0].namespace` | string | | `"jx"` | -| `foghorn.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods | `{}` | -| `foghorn.image.pullPolicy` | string | Template for computing the foghorn controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `foghorn.image.repository` | string | Template for computing the foghorn controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-foghorn"` | -| `foghorn.image.tag` | string | Template for computing the foghorn controller docker image tag | `"{{ .Values.image.tag }}"` | -| `foghorn.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `foghorn.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods | `{}` | -| `foghorn.replicaCount` | int | Number of replicas | `1` | -| `foghorn.resources.limits` | object | Resource limits applied to the foghorn pods | `{"cpu":"100m","memory":"256Mi"}` | -| `foghorn.resources.requests` | object | Resource requests applied to the foghorn pods | `{"cpu":"80m","memory":"128Mi"}` | -| `foghorn.terminationGracePeriodSeconds` | int | Termination grace period for foghorn pods | `180` | -| `foghorn.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods | `[]` | -| `gcJobs.backoffLimit` | int | Set the backoff limit for failed cronJobs | `6` | -| `gcJobs.concurrencyPolicy` | string | Drives the job's concurrency policy | `"Forbid"` | -| `gcJobs.failedJobsHistoryLimit` | int | Drives the failed jobs history limit | `1` | -| `gcJobs.image.pullPolicy` | string | Template for computing the gc job docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `gcJobs.image.repository` | string | Template for computing the gc job docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-gc-jobs"` | -| `gcJobs.image.tag` | string | Template for computing the gc job docker image tag | `"{{ .Values.image.tag }}"` | -| `gcJobs.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `gcJobs.maxAge` | string | Max age from which `LighthouseJob`s will be deleted | `"168h"` | -| `gcJobs.schedule` | string | Cron expression to periodically delete `LighthouseJob`s | `"0/30 * * * *"` | -| `gcJobs.successfulJobsHistoryLimit` | int | Drives the successful jobs history limit | `3` | -| `git.kind` | string | Git SCM provider (`github`, `gitlab`, `stash`) | `"github"` | -| `git.server` | string | Git server URL | `""` | -| `githubApp.enabled` | bool | Enables GitHub app authentication | `false` | -| `githubApp.username` | string | GitHub app user name | `"jenkins-x[bot]"` | -| `hmacSecretName` | string | Existing hmac secret to use for webhooks | `""` | -| `hmacToken` | string | Secret used for webhooks | `""` | -| `hmacTokenEnabled` | bool | Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud | `true` | -| `image.parentRepository` | string | Docker registry to pull images from | `"ghcr.io/jenkins-x"` | -| `image.pullPolicy` | string | Image pull policy | `"IfNotPresent"` | -| `image.tag` | string | Docker images tag the following tag is latest on the main branch, it's a specific version on a git tag | `"latest"` | -| `jenkinscontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | -| `jenkinscontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `jenkinscontroller.image.repository` | string | Template for computing the Jenkins controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller"` | -| `jenkinscontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | -| `jenkinscontroller.jenkinsToken` | string | The token for authenticating the Jenkins user | `nil` | -| `jenkinscontroller.jenkinsURL` | string | The URL of the Jenkins instance | `nil` | -| `jenkinscontroller.jenkinsUser` | string | The username for the Jenkins user | `nil` | -| `jenkinscontroller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `jenkinscontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | -| `jenkinscontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | -| `jenkinscontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | -| `jenkinscontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | -| `jenkinscontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | -| `jenkinscontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | -| `jenkinscontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | -| `keeper.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods | `{}` | -| `keeper.datadog.enabled` | string | Enables datadog | `"true"` | -| `keeper.env` | object | Lets you define keeper specific environment variables | `{}` | -| `keeper.image.pullPolicy` | string | Template for computing the keeper controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `keeper.image.repository` | string | Template for computing the keeper controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-keeper"` | -| `keeper.image.tag` | string | Template for computing the keeper controller docker image tag | `"{{ .Values.image.tag }}"` | -| `keeper.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `keeper.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `keeper.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods | `{}` | -| `keeper.podAnnotations` | object | Annotations applied to the keeper pods | `{}` | -| `keeper.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `keeper.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `keeper.replicaCount` | int | Number of replicas | `1` | -| `keeper.resources.limits` | object | Resource limits applied to the keeper pods | `{"cpu":"400m","memory":"512Mi"}` | -| `keeper.resources.requests` | object | Resource requests applied to the keeper pods | `{"cpu":"100m","memory":"128Mi"}` | -| `keeper.service` | object | Service settings for the keeper controller | `{"externalPort":80,"internalPort":8888,"type":"ClusterIP"}` | -| `keeper.statusContextLabel` | string | Label used to report status to git provider | `"Lighthouse Merge Status"` | -| `keeper.terminationGracePeriodSeconds` | int | Termination grace period for keeper pods | `30` | -| `keeper.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods | `[]` | -| `lighthouseJobNamespace` | string | Namespace where `LighthouseJob`s and `Pod`s are created | Deployment namespace | -| `logFormat` | string | Log format either json or stackdriver | `"json"` | -| `logService` | string | The name of the service registered with logging | `""` | -| `logStackSkip` | string | Comma separated stack frames to skip from the log | `""` | -| `oauthSecretName` | string | Existing Git token secret | `""` | -| `oauthToken` | string | Git token (used when GitHub app authentication is not enabled) | `""` | -| `poller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods | `{}` | -| `poller.contextMatchPattern` | string | Regex pattern to use to match commit status context | `""` | -| `poller.datadog.enabled` | string | Enables datadog | `"true"` | -| `poller.enabled` | bool | Whether to enable or disable the poller component | `false` | -| `poller.env` | object | Lets you define poller specific environment variables | `{"POLL_HOOK_ENDPOINT":"http://hook/hook/poll","POLL_PERIOD":"20s"}` | -| `poller.image.pullPolicy` | string | Template for computing the poller controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `poller.image.repository` | string | Template for computing the poller controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-poller"` | -| `poller.image.tag` | string | Template for computing the poller controller docker image tag | `"{{ .Values.image.tag }}"` | -| `poller.internalPort` | int | | `8888` | -| `poller.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `poller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `poller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods | `{}` | -| `poller.podAnnotations` | object | Annotations applied to the poller pods | `{}` | -| `poller.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `poller.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `poller.replicaCount` | int | Number of replicas | `1` | -| `poller.requireReleaseSuccess` | bool | Keep polling releases until the most recent commit status is successful | `false` | -| `poller.resources.limits` | object | Resource limits applied to the poller pods | `{"cpu":"400m","memory":"512Mi"}` | -| `poller.resources.requests` | object | Resource requests applied to the poller pods | `{"cpu":"100m","memory":"128Mi"}` | -| `poller.terminationGracePeriodSeconds` | int | Termination grace period for poller pods | `30` | -| `poller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods | `[]` | -| `scope` | string | set scope to either `cluster` or `namespace` for permissions | `cluster` | -| `tektoncontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | -| `tektoncontroller.dashboardTemplate` | string | Go template expression for URLs in the dashboard if not using Tekton dashboard | `""` | -| `tektoncontroller.dashboardURL` | string | the dashboard URL (e.g. Tekton dashboard) | `""` | -| `tektoncontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `tektoncontroller.image.repository` | string | Template for computing the tekton controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-tekton-controller"` | -| `tektoncontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | -| `tektoncontroller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `tektoncontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | -| `tektoncontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | -| `tektoncontroller.replicaCount` | int | Number of replicas | `1` | -| `tektoncontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | -| `tektoncontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | -| `tektoncontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | -| `tektoncontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | -| `tektoncontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | -| `user` | string | Git user name (used when GitHub app authentication is not enabled) | `""` | -| `webhooks.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods | `{}` | -| `webhooks.customDeploymentTriggerCommand` | string | deployments can configure the ability to allow custom lighthouse triggers using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing | `""` | -| `webhooks.image.pullPolicy` | string | Template for computing the webhooks controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `webhooks.image.repository` | string | Template for computing the webhooks controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-webhooks"` | -| `webhooks.image.tag` | string | Template for computing the webhooks controller docker image tag | `"{{ .Values.image.tag }}"` | -| `webhooks.ingress.annotations` | object | Webhooks ingress annotations | `{}` | -| `webhooks.ingress.enabled` | bool | Enable webhooks ingress | `false` | -| `webhooks.ingress.hosts` | list | Webhooks ingress host names | `[]` | -| `webhooks.ingress.ingressClassName` | string | Webhooks ingress ingressClassName | `nil` | -| `webhooks.labels` | object | allow optional labels to be added to the webhook deployment | `{}` | -| `webhooks.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":60,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `webhooks.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `webhooks.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods | `{}` | -| `webhooks.podAnnotations` | object | Annotations applied to the webhooks pods | `{}` | -| `webhooks.podLabels` | object | | `{}` | -| `webhooks.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `webhooks.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `webhooks.replicaCount` | int | Number of replicas | `1` | -| `webhooks.resources.limits` | object | Resource limits applied to the webhooks pods | `{"cpu":"100m","memory":"512Mi"}` | -| `webhooks.resources.requests` | object | Resource requests applied to the webhooks pods | `{"cpu":"80m","memory":"128Mi"}` | -| `webhooks.service` | object | Service settings for the webhooks controller | `{"annotations":{},"externalPort":80,"internalPort":8080,"type":"ClusterIP"}` | -| `webhooks.serviceName` | string | Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out | `"hook"` | -| `webhooks.terminationGracePeriodSeconds` | int | Termination grace period for webhooks pods | `180` | -| `webhooks.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods | `[]` | +| Key | Type | Description | Default | +|-----|------|-------------|---------| +| `cluster.crds.create` | bool | Create custom resource definitions | `true` | +| `configMaps.config` | string | Raw `config.yaml` content | `nil` | +| `configMaps.configUpdater` | object | Settings used to configure the `config-updater` plugin | `{"orgAndRepo":"","path":""}` | +| `configMaps.create` | bool | Enables creation of `config.yaml` and `plugins.yaml` config maps | `false` | +| `configMaps.plugins` | string | Raw `plugins.yaml` content | `nil` | +| `engines.jenkins` | bool | Enables the Jenkins engine | `false` | +| `engines.jx` | bool | Enables the jx engine | `true` | +| `engines.tekton` | bool | Enables the tekton engine | `false` | +| `env` | object | Environment variables | `{"JX_DEFAULT_IMAGE":""}` | +| `externalPlugins[0].name` | string | | `"cd-indicators"` | +| `externalPlugins[0].requiredResources[0].kind` | string | | `"Service"` | +| `externalPlugins[0].requiredResources[0].name` | string | | `"cd-indicators"` | +| `externalPlugins[0].requiredResources[0].namespace` | string | | `"jx"` | +| `externalPlugins[1].name` | string | | `"lighthouse-webui-plugin"` | +| `externalPlugins[1].requiredResources[0].kind` | string | | `"Service"` | +| `externalPlugins[1].requiredResources[0].name` | string | | `"lighthouse-webui-plugin"` | +| `externalPlugins[1].requiredResources[0].namespace` | string | | `"jx"` | +| `foghorn.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods | `{}` | +| `foghorn.image.pullPolicy` | string | Template for computing the foghorn controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `foghorn.image.repository` | string | Template for computing the foghorn controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-foghorn"` | +| `foghorn.image.tag` | string | Template for computing the foghorn controller docker image tag | `"{{ .Values.image.tag }}"` | +| `foghorn.logLevel` | string | | `"info"` | +| `foghorn.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods | `{}` | +| `foghorn.replicaCount` | int | Number of replicas | `1` | +| `foghorn.resources.limits` | object | Resource limits applied to the foghorn pods | `{"cpu":"100m","memory":"256Mi"}` | +| `foghorn.resources.requests` | object | Resource requests applied to the foghorn pods | `{"cpu":"80m","memory":"128Mi"}` | +| `foghorn.terminationGracePeriodSeconds` | int | Termination grace period for foghorn pods | `180` | +| `foghorn.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods | `[]` | +| `gcJobs.backoffLimit` | int | Drives the job's backoff limit | `6` | +| `gcJobs.concurrencyPolicy` | string | Drives the job's concurrency policy | `"Forbid"` | +| `gcJobs.failedJobsHistoryLimit` | int | Drives the failed jobs history limit | `1` | +| `gcJobs.image.pullPolicy` | string | Template for computing the gc job docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `gcJobs.image.repository` | string | Template for computing the gc job docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-gc-jobs"` | +| `gcJobs.image.tag` | string | Template for computing the gc job docker image tag | `"{{ .Values.image.tag }}"` | +| `gcJobs.logLevel` | string | | `"info"` | +| `gcJobs.maxAge` | string | Max age from which `LighthouseJob`s will be deleted | `"168h"` | +| `gcJobs.schedule` | string | Cron expression to periodically delete `LighthouseJob`s | `"0/30 * * * *"` | +| `gcJobs.successfulJobsHistoryLimit` | int | Drives the successful jobs history limit | `3` | +| `git.kind` | string | Git SCM provider (`github`, `gitlab`, `stash`) | `"github"` | +| `git.server` | string | Git server URL | `""` | +| `githubApp.enabled` | bool | Enables GitHub app authentication | `false` | +| `githubApp.username` | string | GitHub app user name | `"jenkins-x[bot]"` | +| `hmacSecretName` | string | Existing hmac secret to use for webhooks | `""` | +| `hmacToken` | string | Secret used for webhooks | `""` | +| `hmacTokenEnabled` | bool | Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud | `true` | +| `hmacTokenVolumeMount` | object | Mount hmac token as a volume instead of using an environment variable Secret reference | `{"enabled":false}` | +| `image.parentRepository` | string | Docker registry to pull images from | `"ghcr.io/jenkins-x"` | +| `image.pullPolicy` | string | Image pull policy | `"IfNotPresent"` | +| `image.tag` | string | Docker images tag the following tag is latest on the main branch, it's a specific version on a git tag | `"latest"` | +| `jenkinscontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | +| `jenkinscontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `jenkinscontroller.image.repository` | string | Template for computing the Jenkins controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller"` | +| `jenkinscontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | +| `jenkinscontroller.jenkinsToken` | string | The token for authenticating the Jenkins user | `nil` | +| `jenkinscontroller.jenkinsURL` | string | The URL of the Jenkins instance | `nil` | +| `jenkinscontroller.jenkinsUser` | string | The username for the Jenkins user | `nil` | +| `jenkinscontroller.logLevel` | string | | `"info"` | +| `jenkinscontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | +| `jenkinscontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | +| `jenkinscontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | +| `jenkinscontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | +| `jenkinscontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | +| `jenkinscontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | +| `jenkinscontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | +| `keeper.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods | `{}` | +| `keeper.datadog.enabled` | string | Enables datadog | `"true"` | +| `keeper.env` | object | Lets you define keeper specific environment variables | `{}` | +| `keeper.image.pullPolicy` | string | Template for computing the keeper controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `keeper.image.repository` | string | Template for computing the keeper controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-keeper"` | +| `keeper.image.tag` | string | Template for computing the keeper controller docker image tag | `"{{ .Values.image.tag }}"` | +| `keeper.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `keeper.logLevel` | string | | `"info"` | +| `keeper.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods | `{}` | +| `keeper.podAnnotations` | object | Annotations applied to the keeper pods | `{}` | +| `keeper.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `keeper.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `keeper.replicaCount` | int | Number of replicas | `1` | +| `keeper.resources.limits` | object | Resource limits applied to the keeper pods | `{"cpu":"400m","memory":"512Mi"}` | +| `keeper.resources.requests` | object | Resource requests applied to the keeper pods | `{"cpu":"100m","memory":"128Mi"}` | +| `keeper.service` | object | Service settings for the keeper controller | `{"externalPort":80,"internalPort":8888,"type":"ClusterIP"}` | +| `keeper.statusContextLabel` | string | Label used to report status to git provider | `"Lighthouse Merge Status"` | +| `keeper.terminationGracePeriodSeconds` | int | Termination grace period for keeper pods | `30` | +| `keeper.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods | `[]` | +| `lighthouseJobNamespace` | string | Namespace where `LighthouseJob`s and `Pod`s are created | Deployment namespace | +| `logFormat` | string | Log format either json or stackdriver | `"json"` | +| `logService` | string | The name of the service registered with logging | `""` | +| `logStackSkip` | string | Comma separated stack frames to skip from the log | `""` | +| `oauthSecretName` | string | Existing Git token secret | `""` | +| `oauthToken` | string | Git token (used when GitHub app authentication is not enabled) | `""` | +| `oauthTokenVolumeMount` | object | Mount Git token as a volume instead of using an environment variable Secret reference (used when GitHub app authentication is not enabled) | `{"enabled":false}` | +| `poller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods | `{}` | +| `poller.contextMatchPattern` | string | Regex pattern to use to match commit status context | `""` | +| `poller.datadog.enabled` | string | Enables datadog | `"true"` | +| `poller.enabled` | bool | Whether to enable or disable the poller component | `false` | +| `poller.env` | object | Lets you define poller specific environment variables | `{"POLL_HOOK_ENDPOINT":"http://hook/hook/poll","POLL_PERIOD":"20s"}` | +| `poller.image.pullPolicy` | string | Template for computing the poller controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `poller.image.repository` | string | Template for computing the poller controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-poller"` | +| `poller.image.tag` | string | Template for computing the poller controller docker image tag | `"{{ .Values.image.tag }}"` | +| `poller.internalPort` | int | | `8888` | +| `poller.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `poller.logLevel` | string | | `"info"` | +| `poller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods | `{}` | +| `poller.podAnnotations` | object | Annotations applied to the poller pods | `{}` | +| `poller.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `poller.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `poller.replicaCount` | int | Number of replicas | `1` | +| `poller.requireReleaseSuccess` | bool | Keep polling releases until the most recent commit status is successful | `false` | +| `poller.resources.limits` | object | Resource limits applied to the poller pods | `{"cpu":"400m","memory":"512Mi"}` | +| `poller.resources.requests` | object | Resource requests applied to the poller pods | `{"cpu":"100m","memory":"128Mi"}` | +| `poller.terminationGracePeriodSeconds` | int | Termination grace period for poller pods | `30` | +| `poller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods | `[]` | +| `scope` | string | limit permissions to namespace privileges | `"cluster"` | +| `tektoncontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | +| `tektoncontroller.dashboardTemplate` | string | Go template expression for URLs in the dashboard if not using Tekton dashboard | `""` | +| `tektoncontroller.dashboardURL` | string | the dashboard URL (e.g. Tekton dashboard) | `""` | +| `tektoncontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `tektoncontroller.image.repository` | string | Template for computing the tekton controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-tekton-controller"` | +| `tektoncontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | +| `tektoncontroller.logLevel` | string | | `"info"` | +| `tektoncontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | +| `tektoncontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | +| `tektoncontroller.replicaCount` | int | Number of replicas | `1` | +| `tektoncontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | +| `tektoncontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | +| `tektoncontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | +| `tektoncontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | +| `tektoncontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | +| `user` | string | Git user name (used when GitHub app authentication is not enabled) | `""` | +| `webhooks.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods | `{}` | +| `webhooks.customDeploymentTriggerCommand` | string | deployments can configure the ability to allow custom lighthouse triggers using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing | `""` | +| `webhooks.image.pullPolicy` | string | Template for computing the webhooks controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `webhooks.image.repository` | string | Template for computing the webhooks controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-webhooks"` | +| `webhooks.image.tag` | string | Template for computing the webhooks controller docker image tag | `"{{ .Values.image.tag }}"` | +| `webhooks.ingress.annotations` | object | Webhooks ingress annotations | `{}` | +| `webhooks.ingress.enabled` | bool | Enable webhooks ingress | `false` | +| `webhooks.ingress.hosts` | list | Webhooks ingress host names | `[]` | +| `webhooks.ingress.ingressClassName` | string | Webhooks ingress ingressClassName | `nil` | +| `webhooks.ingress.tls.enabled` | bool | Enable webhooks ingress tls | `false` | +| `webhooks.ingress.tls.secretName` | string | Specify webhooks ingress tls secretName | `""` | +| `webhooks.labels` | object | allow optional labels to be added to the webhook deployment | `{}` | +| `webhooks.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":60,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `webhooks.logLevel` | string | | `"info"` | +| `webhooks.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods | `{}` | +| `webhooks.podAnnotations` | object | Annotations applied to the webhooks pods | `{}` | +| `webhooks.podLabels` | object | | `{}` | +| `webhooks.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `webhooks.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `webhooks.replicaCount` | int | Number of replicas | `1` | +| `webhooks.resources.limits` | object | Resource limits applied to the webhooks pods | `{"cpu":"100m","memory":"512Mi"}` | +| `webhooks.resources.requests` | object | Resource requests applied to the webhooks pods | `{"cpu":"80m","memory":"128Mi"}` | +| `webhooks.service` | object | Service settings for the webhooks controller | `{"annotations":{},"externalPort":80,"internalPort":8080,"type":"ClusterIP"}` | +| `webhooks.serviceName` | string | Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out | `"hook"` | +| `webhooks.terminationGracePeriodSeconds` | int | Termination grace period for webhooks pods | `180` | +| `webhooks.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods | `[]` | You can look directly at the [values.yaml](./values.yaml) file to look at the options and their default values. diff --git a/charts/lighthouse/lighthouse-1.15.0.tgz b/charts/lighthouse/lighthouse-1.15.0.tgz new file mode 100644 index 0000000000000000000000000000000000000000..33ad845a37856941d607971cbcec337e3410c13a GIT binary patch literal 15774 zcmZ9zQ*8?7DNwmW9Www))oJGO1xw$ZWOv5k&x+t2Rz+y6fK>!1$S7^^g@=9+cS z>n4td!2tPh1JZ!d8A+)yn@GuX$a!+J8?&l2n*!J^wE^55it6ed@)|Z)Ms}v2s>=5K zQf4-`AeTMf&KvFb)E?g}<^5_imD1+5=?3F!%Qm&~2hu+eDD*cnAG%Vr8cETLPyi6n zAZ8zOjrqRtOlTCsiC&s+<{_%wWUwxr`mqwlU8wtpf2b&oD%|-3Q`~R%*gi+S0DU_* z#Md{Iz>hD#JK*=g?k;igSMcEX-rn~1_kiGga8J+RfS{27v!UlTjVEp@6i6Uc8*Qfs#uw`AWxy*};uC_UBwy0u5eJ*bnVBbhgY* zjBPiy$Pm)=#E9YuY`F41o;HgiX2~8JQ-XP-ygwif`ch!VKet>VKI0@sM7nISn#rgm znr>&|o93$Fsn%;IAHG$8X|mBAXK+rAokCJ!C&Cis5$8Kq4{^ z;VqQ2gX1UBP|8$&Jhg{kE#=A+5=x&4wLc5e-@|utco1a=7}Y!sJW-dhioo)Gf!zGo z06&}{x2~tDETz7Ld1x+i`YOETjgSH3KOwGCY<-DPzq>5hGL<=PwPE-a zaqh)U#FtY*r-a1XNpHkuF;?}2#a3jUg2}+MC&Np3&HHUmvB!!d{O8mJI|gG8sOueA zG7HtlbjfZx(I^3rc@&A4rcdO-V7LK>ln-Oeu)+rPa zpiv_cMPoXR{JCsMli>Tf8FGc`t*sJ_EOD@qZ5C{#hhZ%Xx}Ata&f;`Yik+aKRAB13 zs+ID9SE{4CX(}lZQWFqNm%rV5Bu5||q|&G0INN``p(MtZPmm1J%O- z7R*m21QN>oSei^37gA*sL<+_K&{-NrE zBkoU;K?c-)DJOs=WY9Tvr2k-_NV{NiKg%W)3Z~RK3!Tc}RiKB}g8oB07x)5?%fh}U zf|gT+3pf3_?2I0=Uy74GC(7F|Sxm78%7vFq=m2lo431J7avi{O$ZQjcF@w@9{<}4a zilXQmW1A4t$-FzPD7Qhq{YkpW3`pvD)2jFbViU;=qy_jY+GY_c2nK3Qp& z8x7P28$fZ=nu8Z%Ot#L8aKqk=#zA^%Ni8~e{qp?2EBVBv5zI0~dOtkdmJ!MAO6wV! z!5PF|6)I&0W-t~J7&%eqp#TrU89<72c%JM*WNecY!k*C>V`sA*5N0?#a+)|MTcOE& z-IV!@^j$ZM8wBo zyc}-IBceJ+{2pz*sYS_ppaP>C0iKwWvICPfL$TN{L=@>$mc={`0!9l7nfZgNBPin) zY+2Vrl-XmMDeg2S0dmwO>MuEJIRyf$Z#$@6Ld%xeYW$tH6*k^2XF zCS6o@P)#>N(5;Vm3CJ6ediPLv8##1VUiJxODz{X;nUgU;gLag)lr~6yA+sGj1JP| z6~eHTMis-nYYL3t&=dYJMvCBsQ$~0*h4Bi7tp&|ZC85(LiX7pRs+j+MbdUU0pLq3& z7kVq8hlaGg76}g@>Q#NzJ_##sBO+o@s}JJ9Cr+`a^}JtIagmAUomb}>x$YCW_bwe3 z5znX;!KYB7fBOKvAFWljZ5qzZISw-q<2Ya{JxfJ&0V|9?ob%LN5La9fY`%tG@q-0|k*tP0aP^2N=y4bV-0tFhvvy)wgmH=J`K%_oLU}CL4bi z+Zlp*Wc-weRwClCGzAqKbYOG+VVDN^(wCB->p|xlYqUzBultzZQ(AAntUpZJClOMu zf>~}*qU2)P9iTRIl8Te95Jy&MmMU9z&dh*s2OFm_F7-S@51}DxLzN=?lne^s3?vxB zh`rm2(?NO*^-S}#Z==f6}-I~1rUaL!334e8XKZb9Acy6KTNUVXj(KO#tzod_& z>?+cgXz7lO1Q_Uy@g?o(6dU+s+(No*3F8?0&Sl0WdY6p82RD>X(Vh~^p9i9G7zp_m zTa2$zWOHn5=ty03`JNyz0ign1e5>rI*3f-Gib$VT<~$9O)nfj5EP@QsbA$1FErb%C z2>oZQ!a(eZYKx%;{_{+5f6Ihmi6)EPf5>1&HdwM~R>5j}8PN-C+lH#{+RSu4YHJnQ z42U<$O$66w!l_Y-F6SA9zne9Oy@smK!;#@Jm9sb3)L=5qw1)&r}++qy$bZHh= zMLA5UN~+V?EkD;2%!*Z{V1H@!&5J|`QTtgPlp8Ebnt2mR%f`|J>!FHT^SV`!Z70wc z+0sDi1vG^+Ekk4~DrxdTs%vH-{_HXKEJV(@1LC{vybQ^eqh z?8io6!YQXaj-k8Bk;PrNH{U0OPu+(%cbteZvZsKjs&&`bE{liCFozZ7;_?N(L_opl zj_GnA%FCbLZA%KI37b6S4W@AXD4QXkDyPr{AF- z??TDcZF;z|(4ladz6!f>UmoIJI9wR9(?Q0iUS%WAOm3kkA&9pvn)!*~w)*QKnd57* z{`HPDTmTj>Sf$UEoBCYQGdPZ;9Cm^wyqwhRM6wFdyz(f(?a8C)B^|+W@p>RV14&vy z!Z@0(a!)PxEcG=DA)U^p@CX7K0!GaG22lDPH71@cHet28r;T6d5rbYQI*rCuq(dXC zeW&uC((Fs5zYyh2o8rOJBaN5~Fa2+HA_)&pJ*|2F2OEVWdW#f~bxu+m)V$QfL>qCG z`BqbWdAY_fs*@Xm-We_R$fMDiE7biq81cpo3%Mu}2wWAg02>Z+RjF63G19fGCb*Il zs&XnxY$}d>N-1_WXqeENs;GR^Z{W+~_^=ru<*3TC8_Cc)vG42kZnZiYxo%dV_rrl^ zUPc3g{IqB+9yYFI(M=9#cIu65T$>GAY?vSS*^4ayojhTEa2SBU{F30E%sLjDQcaN# zvIXlY*E?sopA%3yMrA&QQSNPsDAG}k8Cc2@LUWT^c6(Bl&zPb~&ioyvq)WsrJ>*!S zt1F&Lb3YU?eK0z8-!iGo@j%^ZRzT4*Dof|J?{%dRyQKf8Ct*o{sIPs=u6;hM)Xlrg zR1D$)N$M+7xqOPb6A~~)^Nb_3X{~nIqN36ljwrldt>`3*f`5@!$_a4?Q0W?d>U@5U zq+L6B(|&rLS3A4Y{o;-SW6@lW6B2ak+r%^p_UTw8)&cpqFT}be(`M48F54*+Uv6TU zDW`riG6D_Y7M)6{#)b;X^0_gpu}{918!>F;wG?3zcV*Qfwg$Cd> zsbd+Fk#CNaG{S9z*uy8ddi-puk8@1$qB$h7SfYtQqrnUFyMZe5^NaEwTdsTwoiD|e z$mKj-4RJ$L+s7*bMR!WTm?UN6it=eNnp+6aVr{l?D5`xKU(48s$0dm$SWmvR`0zfc zkRoHr5Z%D;y5kC3t%r|#+wTtOyIWE=KB6@0>9;2romX5wBK2g7yE0~ zB!m#<6WAKW*o0J<&m2A;p4~`hX?5jW{hqVoWi2FCWNeGFTn@lK$W2^t_)m z%dBO%8)s89m|qot3=1$sH6cH%*TTezxHgjo?;go;Z&+g;;7UuqZv76%aLLSvCny3Pnm7|~h?<}d z?^1(2^1c$123iLy;t>*i9GuXPlbA8&!TOVB6e zu8z}1-=X`lO4}xEMJHBb)j6pzYl)as&!L%4TUpil6A5tE%wcDNtdl#cj|(IkrvX|% z#c=8=GQSQV^$iXTxB>eTDa>_mMps6uyH_XVXkslwb?^!B`N>EsjSZohNxOxR>=Kdo z&KrL1Vm2!LXbSjhIw+$hO)4NWF9x8F+}emBaHK6`&Df${9YK#Wwz> z-aaA{>L71vFOM27aOkimV=_EpuLZ(oR!lHD*x@bgc}tm*#Q{@~v)N{!dl_5yLyhv3 zRZ)+Y$~uxhs(JspwL~fnLNjZQ^mV;)3T* zVjzbd4bODI>+r2$Y;G^Y@zA~GnM3UhobYfvg@1TH_7Os{_%>k!M%nGi#4Gf=oB2-J ztD+K{oUVyOvtfIPG{%mCy>p;QZmG6ireKRB_n+6Fh&r}z(HEar>EbqOUm|MV7s52j z`9ba(WwZXXY|O7kE|j|COGKhtV*;^Uc^HkO1CB(p?RMLxqQD3Dde-?>Kc*v-4&_YcW$H;XyB1op~Q2!sPB)S2V;lgFh9=|qs_ahn2X zVj=DnzNy4s=xFcD2KKN$XS-6(rdXSkhcPJkTH%{6T!{~V-zzBf8957u6}!`zvMu4B ziZc%g7ds(SP2NsS>SFxk*%ZH7SMCWvYHVQ%TwnaJZmy1w`nH7qy@9Ld3x9%=%@DN- zh!thBXl1wvNogJL{dd`>G}XVYf`Ymq*Yuha4|9OpI!ohV-?h!CIf6jwudV$rgRZZE z?Q6Ghd$PX|#3L*hrE(ByE$xveVSGdQZn1RHp>L6^9x(bQ&gP4hU_OCb*HKotPS{k5wK`y+X~2 zGktCduf;Hlgi#`%d%k&1Vm!Cr#A{M49e(iYMUUHx5+8EYXD)hI<8fJ$6vCP5DWlzf z2lLFgSq3|;&yxb727ZNnG5!ttnc)q~#3g3I9fd2-8^$_6-L?Os71{W3)e+N~e^zx3iwEr`lw;o9CiyeW!1Iz<~NQr-uZYYo0n z&L0o5tMdi(`UE>`%#i-px?#o~{~LNx#J&J4FNB?Wx_&JIWjkQ_UiKu>k32koJZwu| z}un(OGuX-?bFx$tjhgsokiAjGHZN*KISL>LWyrt zY_Y;&67q*yXgB{3v2zC$81su6VlbcnYG#T*==&3pN)tC4g9uDSiDMg*v5luO_q@~wjr`CCHJS^7|$y zla7z@vJ{u5xYYG53ufg*Ua&|kAUF6@^TXUAd%BP6=@Rd zf}FSSvRju6z8qDTP4>8?YgGA`EwA<;^uAvaaW07XT|fj4S%f+Mr`LN6zuRKNKe!5YY{} z?S6;+M|p%QGn`h=j-Xp@Yo(>rjQ>^q;$5$e@%PQu(@WHYzq>WCA=6axDG>$qCwPL1 zOp(D@h28s!PdC8`@xs^u7xtDbeXP!|^xH%Rz6ZrUnvC+{md@xMWO}^{WvBy!V}wuQ z0fTwY+y{VfcoVWe7a;s5B9YDwj~yj^>J z#lgCoD_mJuy1JG`?Qc?mmp2+`$REd;1D{bl<=tj97Hvbton&pEqsxHp5INppBc#zIG zX1Dn_F~Ykz@Mf!L;NQYRu?+2~6(0WDjWdgXkC`tk^4Hp@J_q0-FRtN=TnCCAb>aq< zbb9m$RR3!D+CbPe`#4yfa1eiPeNMa^l;UHKDO!a46WcoYe`NPYL}=W!nf>THZBpyzJAG1pwAm_Xv(!(z@;mZK;cx4Stlv`A`*GPC--`@> zWoR7${1<4r7XNdLJ~cVe^G8AG59r}n1L%*BjjtU_m9OucJ$i`ai6?7Ltc*TT-3dEG#<81V zNsc@qT?W`v9`FY91XuucvIaIKHjjPo%nj|tGz7Tf+2v;2_lNWg_XiugX2G!4gT4y9 zu5i>+u^$1e|IHis?+R}GY-?))u1i@>05uv>Ug5{2{g;TQ(F7 z{JX(hL`~PjiC7q8;nA&CQiWo%b8x)U71#smgX^ymf9RB5AGg7{a#%UV*~| zsEBHS_{9=SSqOT3h^T|7%}k4vWU^;d2xKfo)Iyr`cXhVrs_U2N))#W{R=W5{UpfvJEz^hNzO%~nAlX>#7r?T$9V)62L?Vy;jE$oEZo`2%OPFfKR&Twhp_B|nq6g|gdg zvO~XWzfH^+IQ2d zh*EZGiWdZ&t7QZ~MEI}2|$adov6$lN~Vlu+I6(3|72RKNOaZrZC5%AC=M z%a92|6ES3hGo5mnBk@Vp)qP~4h+otqF}cb=ML1Z3A~?mPBQ^1yQ|?BKTr_+bj8ZYM ztwPABSP%?~QMPAs?YP%F<3v179k*_$2cx%wHD;(W-FKZtP_q-}>Y&x>6T?UEFcL#J zDH-ww?Z?1Foa2~PsP?aC2Pw0(id2nBdQQj$|W|WKyWNuGX~a z>4nj@ZJk%CSI08QX8%p)XrR>_Dr2*raJS#q@BMl`+dn*CO-^n~b_V^pUmRY(8%0j1 z0y=xTdOhAGFJ^IYEClmSFQWz?UKif$W9ykFUg)tU1(ja-@U9l6S})S+E3r)~YB;MH zWL2(fKmKMV3vusFP=*$7nn76T4sGoe#0x8hFFMivk<7?@lx_OH9M)vrOJF*nW4huQ zE_Wa+lGZ7B3~HxSX*<(+cN@Ijv``Ov0$9EbOU;p4v>7A13`2=E{K5{&_&R&`OQ1?Q zKJ#D2__oe`H3&}Mc*u_}32(Y)u#Q`?TewgHVKN)l342)EYHVP0|E{C@)TsTcBmW*8 zxaRo&kXQiMeBq(&{dtBd0+?x38&|e9=}>m?r^|Q&Q~G(=BB!&gVYXQ1XI-rQWEi7M z7uUtzx#tAX6fd5bdz!C_TEQ8m9zvB7G&3XZGAi{s;$(}=NtKySCVXn7%Q5|>=9bEy zvrCS>PUDaiGQ6|__ux-fB!3R9K1G0w#wJE<> zcO52yoJlK&TOXWhdA-2@F5K|pSGo2_-+}tAXc!mn~b?-%Yi13%y}L z%bcGuLkj0ewLl2=PRo;GEO+iC7+DJ(v$^kb#GKw+^G@rh&Xhud4Y8pXIsGDV<}JU6 zKpDc`IEZ|PhvdVmc?Mp8LWLk8s@~L0Y%MKOwTvD<9j=L_9p5WIS0ZAEFaw z3{dl-_}Y52o@8ZzFwg^Uy&PJ#H2S)+^JiY00OCt4ntl5xazfR*?rQ{p>de0S{BuQ~ zSRK6}{jLwd-TG#rARo^E99!OBw~(D8mL*z&fH(5Hn;meamGTL2<|YD;p}XX6|5$Gdh3$|p!j2RV>~D{b=J8wVT>lz;_k z&;gnm>=nYvAv!>@h?t9nRm`5|7{-v%iPBN1Fut>v*xO0ibqMzb;ZSwKL!knR!>;U* z_gNa$Ymc@pvIa>c5A9}GrSlfuNS0eM@4U+DDcCC^*%_v{m7XxDI6_n-32zOH^o+8= zG%RMv80Yp3^O377=yJlRiQ%dweH6v#TTf6BU=LCv)!DzxBGJ3s0n`5obA)>eA`+Z8 zmh~r*)UXZ*ZWK>>rffF&uYANdgtC`Erpe7QZ6uqT-?LWhpHv<%X>fKZ;BZ3++gy$J z7&pO@;?bFSDx}{=<7@68E^OHb(g^zKdIE>-A1mCHmIA@u^2gsd*wKGJll(ezYyLjP={CErmcyCi=%28?(0;7A|+| z(R1zeBEpA{67ti_Tp4M!&XLUakN=mYfSJ||P)z6}#=t_NnEqDa>BHR={i=zhI^t^C zVGau(^V$=(<9~baRe53)HnI^+nrkC6!(N*O&0mlAi0SrbYHolplW(x^i;{M+=Wm3S zw?V`CuFK8dGR!7Za6~!HmRru=*nIgABM!N!oN-8qqYn$FiE2OGIjaViZL{v7;PjnY zJ|~+A-qaAd{|%<3IZ)J6%u;;&X8T&6_hI-!jKRIa3afva{zWGeT=+^WxLGtuzL*lI zu-wzAQTajn8b6_AMAb9@w8H?b^*ag9J)ZOr?%68(cE7zp;`rV%R}tP<;_#n04EBIL^hDZs#FOYP3~r}YIv~;4 z3cnXF=Dpw{9MbD)DF|-IA0k8AzY{5(!RLw4NK`u_A=1$6+H*jOmwmCN-zOnravvS} zlYZSbnp@7-vK`&w1J#jGF1&{p=vH3r?0T4TGeo0=sCExazu3#(pJOThQ7AYyQyG-d_vWy z>%ZK3ne})>5dptCQp?aU zs@uzH7Yc&Bp^2O=^)TblfeG9zFlYq+CA7H$awJ*J`V%n-v|6BDzxjdzF)I4Wrpkz{y@Vz58&+eu7fqOB~cY0#gg#Z0B*g^4%F%I z&Q!7CZ+Xqif04<(t>4t{ONF{C@5X11m8VnwJTXlmBiW}(CBrxWx)Cz zTqRM0Z9E2m;Eu%t7>x)VUf=DoWmgSSZB_M}uLx<@@M?${P=PEsE|Zm~Plto_8KWcz ztK5;ccwOuc%s>8g@#8^`J4-R|XOze0XdulZQ)n1H&P3IEvBtC^$s!)mx z7<@Y^Q+O>cbIU7abolxA(Wde5*`aab5+)$aA4-eMr-s*$o>-{tB;Kp3{J(%!Ee+V*=bPXf?;6vv|Mte#xj#qJiqW=F*yQmv zk8TFLr0g%_EJ@VEI1s>;s%Kva?`tffK-(^I%01KJR=sJJD9D7@?if5eap1`U5$I== zHPT=gFbpAgygqngaSd*G(~lGqjU%uNI-EWe?BxktK_{&_$-2SF@#Vd$l@9*NLhf^s zGB@ie`@%K_Ingw|)mHx}V0<{3Ty3D{N1vl^Q4Yhe_=OW_(lpz>9p%*4cUOGk2)P!_7_H`iJ zOGSIt`n}r&ypT5s5P6K{8(yk*Yqv3D4{oLr$xQ?4pwZSLtzjvas8ig>yLc2)gs3RV0ff{vZsh~L4Obno~4uN0DhI41miG{YY53xnTzFXKN{)e9<<2m z2>$Rc^HdYZu=A^DS{JA+yU(3fcG>5np9z>*@FdHu&9Wu z+!`^!7#duYapFbCG%crT=hde4epZ2!Ai^Zue$oxd!yIIBy9anEtzDV86|@PN$71qk zZytO7D?VYod2%xIK>>B$nom}pI#{a^F0EH?PmYD^xnREEz&@L-`pk%bnoHX)PRR89 zM>bW8(Z1OKhQDZ1#}Crqn=Wkz@#KsZWfM=tv&TMP=VV&VF*`C_XCo%6p?NG<{nhBv z<9x7Ci(kce%qp4^PR~SguY|uiIgl-Xb}Xug;fDCIFRiB^N3XF%nRpae+XsDTmaQ)L zaMbrC7EsglPao*EfIjtossC0JqM4q>A7eFJwTq#q-0lk4C8O6lh$UHXeQ7WW|5U5q zyWU_E_|*0-m@SRGyM31PifNbov)N<@QaM`XG+q_Xm?#p}!h52%*OuZ-_i{962>=6T ziFA7IORZ;@hk2c1_J>zurKbOZ1Su=ku1Ww-xq%jufRlL6*2Q(7$ouyv=AzGcUmi?> z>-~=5i<$KI#>PD`wuHUi{agM=d7ng8Ca}8Qx2#g(`{E|`urD~hYvZjivbo*+S8deq zkD>)Gip9Cw4PH}H$c{7hs1;ayb#Cs`De?8$h1_p9SF(cL$UPR{ z8z##gFGhiNPY9n-G?cL5H~RVY>YQbV@b@0o(Y_Ggw8TFQ{+%Q9y(q9JoKx!4%J*|N zJozDd_uLf5~3G`<3<-DvTXTXS0|YuuYDiw0cqG9GLqYSRtC&~1)IcsZcpP9|pum!6hbr;>VviD~yg$H9M9>n5D+>;(O zMZ%kY^S_(a5ni>?m@kI~$&C|EW3uIjUzgIW%^ld*5w!X1Rvxk$FiVb+w6T*fnB0qT z35kPVwGraTq5?t270eYO&Lr724MupCqiS)l1Qc4|BoWqG*>A;0)*7wIY|o{ z>r2r$Ho1cgE^>9R%80)Sw>htFrQ#-(<-+31mIeJ8=1~PKoejYvD)m})gcpf>Nx1%D zZD5W}vv27QGMMLog<~AHJEZf-WiRzn(P^!#ts_iaGI%g3e|>Cu!ql~P^Sg1idC6yWAWj9Up;EHzetzBrM#Xiaf0vIsw z3%X!I$SfXq$#hCi4M^RX5_4Gq{1vXZNHti}C~(4-mNwc{!vH${N^SPyCswE`s*ht= z3-wC7*h;ntqlqjl%43yk03+wUNe6a#`QAUwmoH9MyWhlb=d2N?U$?3S`(6VXIXgSB zFqPKxb_NpT`Nv@YRiqa%z7h?iKAoMB$n>{&$C!eb)>Fg-`9XYXsAvYioHQV9l*FUV z9Q~ptV6Q9Yxu9p>@oM4Lf>qWl-u?%vm2#YQ%zwO%X;w{w&&zprN=rmYM38%qoTJA- z@u7KAAJ>fj<89F!pxovVyo1^v{~tMsdZ&DoDZCn(xD$vvO89UL8Q@4@Jp3>JS;Y7U znOYtD&?V?2m?Tu&c^QRkUhw~nMIW$r4rppO>8IL8ETRnpLqQR)AtR5FXj0WJ9fRHJ z#l&yMnK05*)fyn(6cdgj^UFJWf1Z9`nBU!6a>A*JmO}#P^xYTjXo*2~@T}PXNBwlz z+Q)nbd0<}2Aun{R)pSvQ#!~Z9r!E6J@%L$t=`{>k zcLLwlWi>aVlb`GPPhZNfa-T8eVSDV6H~zVw8^D96|gQG-bHv{g3 z*{9v{_1mA5P+*CQ-nJ8{IuwStRR%*zo+B_BPQzd|MDB&3{%*g)UD$H zL3Ex!e!R$D|HU|W4-Do(2daxzjc^1Aa;nEM`@%$*jk^4|5TC9N_XMQTtWMaZbU$zU zFCto<(b_?7+>}=m46IY8*IH)yLfs{W^!Am29`yduI6QiL``#A6YYdfsHjGNI zS+xK9OBeQ;2U-VCcy#{%@5Xo(TjgyOgE?XvAZk@GKXt2njfkVZ{W(uQ)?P)1+b~^1yHwaVeG@iP{(wSBlUkm&risGkOXFO z5R#wg(l8_SRFz$hs2h)p6oRQjv> zEbM8vm7K`Hn``arrrCLZbA3?Qzg!PtY=sEPZT+pl>T0imxjBpBkmmtir^nj-(doj4 zaPC|`Br#tcQPp>%(M;;OTmby01zzbP@fNPy3aQ9CrN)|`-We?cOHW^krnNl{@)>?% ze86xO4NAr%9bKB%0sBXBH%XBr6NPVzR+I{#X4=KoUiG)aMysSfvO$m6lU4cjjmbYq zW)VM8CPfN+zy!_FWz@{&d($P&UIiU#z=#IAD#pX8+Wv5=P0n+h9FN3-^Z_elfakDK z7HmEXFb4X^Cy4s>j}!z9X71(IE5@wLY}88`eiO<>X>H*q);U8t(UJCWXa&np zBx#fiw9NbYII$cJ&;PHB#*FZXo}rxhl^CE2pMcIRlZ}Afvpy8&*zW5bf9lCjmrhF- zWrFv$X3CkHx1`ps1qqkclRGjE#X!|CqpojC>9?eB2mPKJ&+Kbu^~HYFbQ9$!K<<|R za9d30;6%PhDd1YA*rPV$MC|VhJc~mVV7tQT=v}K=%dD+HI7N}=n^oB9*{P^YM$YIT zUAkD+g`zH9QoDEwh{d@7^o@X8G$dXS`0()ZdU`$>HKm z zRr}Mm8CMPV@=2)CEr4kQ>fd=3kstM&{kKy6#2i0&TB#hnv9gY(G@1CB%_s_hZ;%vj zNjq=kbADH5bzML&v2I{qx{^b{G#NE;iF49K|CTDVtJN%$d}Orz0btIjsX}tBFY+VD zfYz(At~EeYUi!|4DSmJ9jm9IcQ@9P{d)Vhy@tnFdsB?I@8t+^kqfREhadocm&g2x0 zaN0EOUT;}Z)=SHrQgF@D8g}t#V*vQr0pZ^n(qm->n`*{b{|N96wBPc%t7)i$oiOve z?Hn{&A>9L;jjp;{#Oj@j`)84um0zZjtZdzZtgC*1pSBS^m}xy^s_IV`1W>|AJJ`4L z(wu2E>NBx^rT~~aVujH51kuh4Cj;Rg>&c<1I)fLv#AwHQ?N^A=RZjmeo?r9-@ci%1 z;EO1pyzx&_X39ft9brY(n3A{Co*GhwO!fTIl5=g`oE{cZL$4cDw0mh~EnUw=>%~-# ztG=fw=G2$1-?6_jL$Sg8f_5QIQ~zNwGIS4ivSZU9K%5XB@-fL+r>XuwTev@*k86;i zOpUrO<38+Ij?v`Fg4ZRExXiDX;XBlsgcv5$I%b3HW;ZhVy9yUjX|Ar;!E zddbFFeAONsbUEp;eleOX-Hk)s;FE&*_6-Fw9G?+ETCY+O{T9vD?4)q%b-3;xK7)lh zLR=X*GQBkqP@Fs{C%f@lPQYA6vC~DhoNsu2F7|obL)(aR5>&h7xQ7#lzCLopO?^GF zV~k4t%(A7`mCdu(X#m(~&12ut%CqY3;W>ZqNij?Ruipl6H%-tM{qzg^7B+3&&{D{7 z#eS?IjglrPM$86U34?;PL?u8`&!ZbfxJoENEY}z=O6Jmn5WiFf+x&xk^~pZInTb}p z66<9JNE(@9A=-gu6vaoh4Nm>T>{bi%Fws%i&wu+$E1t>J<|n9L+7jeMp=XTT?`VNQ zC9{o3knt#B85W$Z74iSgoEW*9$&q}1aaf??wKIO-OzwXncYW+Hb9`?=B@qI#N4UB^ zXF6|$At43QK1`dKdlks}xr-o}m_rtXx&k{9%FcU1uZGHlKeDpbeu3kRpmSN6rTdZ! zx+n4+g!}FE|pdM&0S+iS?F6 zBuSPEA+SrKRm8dIY_ZL z>CJG2Q_Nc2TBBi0D!1%T@7TZ*1cXcj7*{AICzpUfqJ#z5uM7&~A7KPyVwNKHC^X41 zZ~waB=tuYzE7EjI(#ovQU!>KT$c`SO%-+*GR{W%`*-B9TILVN^ZP$E8a7Y5K*46^~7Ofc$GM0rX3@PVm1a!m`*i zL67npev+!%IWIbHg6^48fD3qw#f&c%E!uOZ-=@C$ z54PVv-TBLY`un@%1D(k$L4g|X1p=YsyEL~7 zr%P2T&<+01-~HtosqID3hJwHTQ-eRGHE2iu_QM;-uP;IQDC<$&pN>i5-L0)kO$yV7 z`fyr$@(wu62=Q<3v+QSIUt15*h};tWe&;;8ACRfHQ<=~vztJqeaVQ1!|F0Drc!3Tg K4)P8P@_zs}TK4_` literal 0 HcmV?d00001 diff --git a/charts/lighthouse/values.yaml b/charts/lighthouse/values.yaml index 54f43949c..d7fb6a038 100644 --- a/charts/lighthouse/values.yaml +++ b/charts/lighthouse/values.yaml @@ -1,80 +1,57 @@ git: # git.kind -- Git SCM provider (`github`, `gitlab`, `stash`) kind: github - # git.server -- Git server URL server: "" - # lighthouseJobNamespace -- Namespace where `LighthouseJob`s and `Pod`s are created # @default -- Deployment namespace lighthouseJobNamespace: "" - githubApp: # githubApp.enabled -- Enables GitHub app authentication enabled: false - # githubApp.username -- GitHub app user name - username: "jenkins-x[bot]" - + username: "jenkins-x[bot]" # user -- Git user name (used when GitHub app authentication is not enabled) user: "" - # oauthToken -- Git token (used when GitHub app authentication is not enabled) oauthToken: "" - # oauthSecretName -- Existing Git token secret oauthSecretName: "" - # oauthTokenVolumeMount -- Mount Git token as a volume instead of using an environment variable Secret reference (used when GitHub app authentication is not enabled) oauthTokenVolumeMount: enabled: false - # hmacToken -- Secret used for webhooks hmacToken: "" - # hmacSecretName -- Existing hmac secret to use for webhooks hmacSecretName: "" - # hmacTokenEnabled -- Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud hmacTokenEnabled: true - # hmacTokenVolumeMount -- Mount hmac token as a volume instead of using an environment variable Secret reference hmacTokenVolumeMount: enabled: false - # logFormat -- Log format either json or stackdriver logFormat: "json" - # logService -- The name of the service registered with logging logService: "" - # logStackSkip -- Comma separated stack frames to skip from the log logStackSkip: "" - # scope -- limit permissions to namespace privileges scope: "cluster" - cluster: crds: # cluster.crds.create -- Create custom resource definitions create: true - image: # image.parentRepository -- Docker registry to pull images from parentRepository: ghcr.io/jenkins-x - # image.tag -- Docker images tag # the following tag is latest on the main branch, it's a specific version on a git tag - tag: latest - + tag: 1.15.0 # image.pullPolicy -- Image pull policy pullPolicy: IfNotPresent - # env -- Environment variables env: JX_DEFAULT_IMAGE: "" - - externalPlugins: - name: cd-indicators requiredResources: @@ -86,392 +63,287 @@ externalPlugins: - kind: Service namespace: jx name: lighthouse-webui-plugin - gcJobs: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # gcJobs.maxAge -- Max age from which `LighthouseJob`s will be deleted maxAge: 168h - # gcJobs.schedule -- Cron expression to periodically delete `LighthouseJob`s schedule: "0/30 * * * *" - # gcJobs.failedJobsHistoryLimit -- Drives the failed jobs history limit failedJobsHistoryLimit: 1 - # gcJobs.successfulJobsHistoryLimit -- Drives the successful jobs history limit successfulJobsHistoryLimit: 3 - # gcJobs.concurrencyPolicy -- Drives the job's concurrency policy concurrencyPolicy: Forbid - # gcJobs.backoffLimit -- Drives the job's backoff limit backoffLimit: 6 - image: # gcJobs.image.repository -- Template for computing the gc job docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-gc-jobs" - # gcJobs.image.tag -- Template for computing the gc job docker image tag tag: "{{ .Values.image.tag }}" - # gcJobs.image.pullPolicy -- Template for computing the gc job docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - webhooks: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # webhooks.replicaCount -- Number of replicas replicaCount: 1 - # webhooks.terminationGracePeriodSeconds -- Termination grace period for webhooks pods terminationGracePeriodSeconds: 180 - image: # webhooks.image.repository -- Template for computing the webhooks controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-webhooks" - # webhooks.image.tag -- Template for computing the webhooks controller docker image tag tag: "{{ .Values.image.tag }}" - # webhooks.image.pullPolicy -- Template for computing the webhooks controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - - # webhooks.labels -- allow optional labels to be added to the webhook deployment labels: {} podLabels: {} - # webhooks.podAnnotations -- Annotations applied to the webhooks pods podAnnotations: {} - # webhooks.serviceName -- Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out serviceName: hook - # webhooks.service -- Service settings for the webhooks controller service: type: ClusterIP externalPort: 80 internalPort: 8080 annotations: {} - resources: # webhooks.resources.limits -- Resource limits applied to the webhooks pods limits: cpu: 100m # may require more memory to perform the initial 'git clone' cmd for big repositories memory: 512Mi - # webhooks.resources.requests -- Resource requests applied to the webhooks pods requests: cpu: 80m memory: 128Mi - # webhooks.probe -- Liveness and readiness probes settings probe: path: / - # webhooks.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 60 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # webhooks.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # webhooks.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods nodeSelector: {} - # webhooks.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods affinity: {} - # webhooks.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods tolerations: [] - ingress: # webhooks.ingress.enabled -- Enable webhooks ingress enabled: false - # webhooks.ingress.annotations -- Webhooks ingress annotations annotations: {} - # webhooks.ingress.ingressClassName -- Webhooks ingress ingressClassName ingressClassName: null - # webhooks.ingress.hosts -- Webhooks ingress host names hosts: [] - tls: # webhooks.ingress.tls.enabled -- Enable webhooks ingress tls enabled: false # webhooks.ingress.tls.secretName -- Specify webhooks ingress tls secretName secretName: "" - # webhooks.customDeploymentTriggerCommand -- deployments can configure the ability to allow custom lighthouse triggers # using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify # `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing customDeploymentTriggerCommand: "" - foghorn: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # foghorn.replicaCount -- Number of replicas replicaCount: 1 - # foghorn.terminationGracePeriodSeconds -- Termination grace period for foghorn pods terminationGracePeriodSeconds: 180 - image: # foghorn.image.repository -- Template for computing the foghorn controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-foghorn" - # foghorn.image.tag -- Template for computing the foghorn controller docker image tag tag: "{{ .Values.image.tag }}" - # foghorn.image.pullPolicy -- Template for computing the foghorn controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - resources: # foghorn.resources.limits -- Resource limits applied to the foghorn pods limits: cpu: 100m memory: 256Mi - # foghorn.resources.requests -- Resource requests applied to the foghorn pods requests: cpu: 80m memory: 128Mi - # foghorn.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods nodeSelector: {} - # foghorn.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods affinity: {} - # foghorn.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods tolerations: [] - - tektoncontroller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # tektoncontroller.dashboardURL -- the dashboard URL (e.g. Tekton dashboard) dashboardURL: '' # tektoncontroller.dashboardTemplate -- Go template expression for URLs in the dashboard if not using Tekton dashboard dashboardTemplate: '' - # tektoncontroller.replicaCount -- Number of replicas replicaCount: 1 - # tektoncontroller.terminationGracePeriodSeconds -- Termination grace period for tekton controller pods terminationGracePeriodSeconds: 180 - image: # tektoncontroller.image.repository -- Template for computing the tekton controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-tekton-controller" - # tektoncontroller.image.tag -- Template for computing the tekton controller docker image tag tag: "{{ .Values.image.tag }}" - # tektoncontroller.image.pullPolicy -- Template for computing the tekton controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # tektoncontroller.podAnnotations -- Annotations applied to the tekton controller pods podAnnotations: {} - # tektoncontroller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods nodeSelector: {} - # tektoncontroller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods affinity: {} - # tektoncontroller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods tolerations: [] - resources: # tektoncontroller.resources.limits -- Resource limits applied to the tekton controller pods limits: cpu: 100m memory: 256Mi - # tektoncontroller.resources.requests -- Resource requests applied to the tekton controller pods requests: cpu: 80m memory: 128Mi - # tektoncontroller.service -- Service settings for the tekton controller service: annotations: {} - jenkinscontroller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # jenkinscontroller.jenkinsURL -- The URL of the Jenkins instance jenkinsURL: - # jenkinscontroller.jenkinsUser -- The username for the Jenkins user jenkinsUser: - # jenkinscontroller.jenkinsToken -- The token for authenticating the Jenkins user jenkinsToken: - # jenkinscontroller.terminationGracePeriodSeconds -- Termination grace period for tekton controller pods terminationGracePeriodSeconds: 180 - image: # jenkinscontroller.image.repository -- Template for computing the Jenkins controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller" - # jenkinscontroller.image.tag -- Template for computing the tekton controller docker image tag tag: "{{ .Values.image.tag }}" - # jenkinscontroller.image.pullPolicy -- Template for computing the tekton controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # jenkinscontroller.podAnnotations -- Annotations applied to the tekton controller pods podAnnotations: {} - # jenkinscontroller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods nodeSelector: {} - # jenkinscontroller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods affinity: {} - # jenkinscontroller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods tolerations: [] - resources: # jenkinscontroller.resources.limits -- Resource limits applied to the tekton controller pods limits: cpu: 100m memory: 256Mi - # jenkinscontroller.resources.requests -- Resource requests applied to the tekton controller pods requests: cpu: 80m memory: 128Mi - # jenkinscontroller.service -- Service settings for the tekton controller service: annotations: {} - keeper: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # keeper.statusContextLabel -- Label used to report status to git provider statusContextLabel: "Lighthouse Merge Status" - # keeper.replicaCount -- Number of replicas replicaCount: 1 - # keeper.terminationGracePeriodSeconds -- Termination grace period for keeper pods terminationGracePeriodSeconds: 30 - image: # keeper.image.repository -- Template for computing the keeper controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-keeper" - # keeper.image.tag -- Template for computing the keeper controller docker image tag tag: "{{ .Values.image.tag }}" - # keeper.image.pullPolicy -- Template for computing the keeper controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # keeper.podAnnotations -- Annotations applied to the keeper pods podAnnotations: {} - # keeper.env -- Lets you define keeper specific environment variables env: {} - # keeper.service -- Service settings for the keeper controller service: type: ClusterIP externalPort: 80 internalPort: 8888 - resources: # keeper.resources.limits -- Resource limits applied to the keeper pods limits: cpu: 400m memory: 512Mi - # keeper.resources.requests -- Resource requests applied to the keeper pods requests: cpu: 100m memory: 128Mi - # keeper.probe -- Liveness and readiness probes settings probe: path: / - # keeper.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 120 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # keeper.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - datadog: # keeper.datadog.enabled -- Enables datadog enabled: "true" - # keeper.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods nodeSelector: {} - # keeper.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods affinity: {} - # keeper.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods tolerations: [] - poller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # poller.enabled -- Whether to enable or disable the poller component enabled: false - # poller.replicaCount -- Number of replicas replicaCount: 1 - # poller.terminationGracePeriodSeconds -- Termination grace period for poller pods terminationGracePeriodSeconds: 30 - image: # poller.image.repository -- Template for computing the poller controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-poller" - # poller.image.tag -- Template for computing the poller controller docker image tag tag: "{{ .Values.image.tag }}" - # poller.image.pullPolicy -- Template for computing the poller controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # poller.podAnnotations -- Annotations applied to the poller pods podAnnotations: {} - # poller.env -- Lets you define poller specific environment variables env: # poller.env.POLL_PERIOD the default time period between polling releases and pull requests POLL_PERIOD: 20s - # poller.env.POLL_RELEASE_PERIOD the time period between polling releases # POLL_RELEASE_PERIOD: 20s @@ -480,77 +352,58 @@ poller: # poller.env.POLL_HOOK_ENDPOINT the hook service endpoint to post webhooks to POLL_HOOK_ENDPOINT: http://hook/hook/poll - # poller.contextMatchPattern -- Regex pattern to use to match commit status context contextMatchPattern: "" - # poller.requireReleaseSuccess -- Keep polling releases until the most recent commit status is successful requireReleaseSuccess: false - resources: # poller.resources.limits -- Resource limits applied to the poller pods limits: cpu: 400m memory: 512Mi - # poller.resources.requests -- Resource requests applied to the poller pods requests: cpu: 100m memory: 128Mi - # poller.probe -- Liveness and readiness probes settings probe: path: / - # keeper.internalPort -- The internal port used to view metrics etc internalPort: 8888 - # poller.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 120 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # poller.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - datadog: # poller.datadog.enabled -- Enables datadog enabled: "true" - # poller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods nodeSelector: {} - # poller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods affinity: {} - # poller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods tolerations: [] - engines: # engines.jx -- Enables the jx engine jx: true - # engines.tekton -- Enables the tekton engine tekton: false - # engines.jenkins -- Enables the Jenkins engine jenkins: false - configMaps: # configMaps.create -- Enables creation of `config.yaml` and `plugins.yaml` config maps create: false - # configMaps.config -- Raw `config.yaml` content config: null - # configMaps.plugins -- Raw `plugins.yaml` content plugins: null - # configMaps.configUpdater -- Settings used to configure the `config-updater` plugin configUpdater: orgAndRepo: ""