Tekton Config
TektonConfig custom resource is the top most component of the operator which allows user to install and customize all other components at a single place.
Operator provides support for installing and managing following operator components through TektonConfig
:
Other than the above components depending on the platform operator also provides support for
- On both Kubernetes and OpenShift
- On Kubernetes
- On OpenShift
The TektonConfig CR provides the following features
apiVersion: operator.tekton.dev/v1alpha1
kind: TektonConfig
metadata:
name: config
spec:
targetNamespace: tekton-pipelines
profile: all
config:
nodeSelector: <>
tolerations: []
priorityClassName: system-cluster-critical
chain:
disabled: false
pipeline:
disable-affinity-assistant: false
disable-creds-init: false
disable-home-env-overwrite: true
disable-working-directory-overwrite: true
enable-api-fields: stable
enable-custom-tasks: false
enable-tekton-oci-bundles: false
metrics.pipelinerun.duration-type: histogram
metrics.pipelinerun.level: pipelinerun
metrics.taskrun.duration-type: histogram
metrics.taskrun.level: taskrun
require-git-ssh-secret-known-hosts: false
running-in-environment-with-injected-sidecars: true
trusted-resources-verification-no-match-policy: ignore
performance:
disable-ha: false
buckets: 1
replicas: 1
threads-per-controller: 2
kube-api-qps: 5.0
kube-api-burst: 10
options:
disabled: false
configMaps: {}
deployments: {}
pruner:
disabled: false
schedule: "0 8 * * *"
resources:
- taskrun
- pipelinerun
keep: 3
# keep-since: 1440
# NOTE: you can use either "keep" or "keep-since", not both
prune-per-resource: true
hub:
params:
- name: enable-devconsole-integration
value: "true"
options:
disabled: false
configMaps: {}
deployments: {}
dashboard:
readonly: true
options:
disabled: false
configMaps: {}
deployments: {}
platforms:
openshift:
pipelinesAsCode:
enable: true
settings:
application-name: Pipelines as Code CI
auto-configure-new-github-repo: "false"
bitbucket-cloud-check-source-ip: "true"
custom-console-name: ""
custom-console-url: ""
custom-console-url-pr-details: ""
custom-console-url-pr-tasklog: ""
error-detection-from-container-logs: "false"
error-detection-max-number-of-lines: "50"
error-detection-simple-regexp: ^(?P<filename>[^:]*):(?P<line>[0-9]+):(?P<column>[0-9]+):([
]*)?(?P<error>.*)
error-log-snippet: "true"
hub-catalog-name: tekton
hub-url: https://api.hub.tekton.dev/v1
remote-tasks: "true"
secret-auto-create: "true"
secret-github-app-token-scoped: "true"
options:
disabled: false
configMaps: {}
deployments: {}
Look for the particular section to understand a particular field in the spec.
Target Namespace
This allows user to choose a namespace to install the Tekton Components such as pipelines, triggers.
By default, namespace would be tekton-pipelines
for Kubernetes and openshift-pipelines
for OpenShift.
Note: Namespace openshift-operators
is not allowed in OpenShift
as a targetNamespace
.
Profile
This allows user to choose which all components to install on the cluster. There are 3 profiles available:
all
: This profile will install all components (TektonPipeline, TektonTrigger and TektonChain)basic
: This profile will install only TektonPipeline, TektonTrigger and TektonChain componentlite
: This profile will install only TektonPipeline component
On Kubernetes, all
profile will install TektonDashboard
and on OpenShift TektonAddon
will be installed.
Config
Config provides fields to configure deployments created by the Operator. This provides following fields:
User can pass the required fields and this would be passed to all Operator components which will get added in all deployments created by Operator.
Example:
config:
nodeSelector:
key: value
tolerations:
- key: "key1"
operator: "Equal"
value: "value1"
effect: "NoSchedule"
priorityClassName: system-node-critical
This is an Optional
section.
NOTE: If spec.config.priorityClassName
is used, then the required priorityClass
is
expected to be created by the user to get the Tekton resources pods in running state
Pipeline
Pipeline section allows user to customize the Tekton pipeline features. This allow user to customize the values in configmaps.
Refer to properties section in TektonPipeline for available options.
Some of the fields have default values, so operator will add them if user haven’t passed in CR. Other fields which doesn’t have default values won’t be passed unless user specify them in CR. User can find those here.
Example:
pipeline:
disable-affinity-assistant: false
disable-creds-init: false
disable-home-env-overwrite: true
disable-working-directory-overwrite: true
enable-api-fields: stable
enable-custom-tasks: false
enable-tekton-oci-bundles: false
metrics.pipelinerun.duration-type: histogram
metrics.pipelinerun.level: pipelinerun
metrics.taskrun.duration-type: histogram
metrics.taskrun.level: taskrun
require-git-ssh-secret-known-hosts: false
running-in-environment-with-injected-sidecars: true
trusted-resources-verification-no-match-policy: ignore
performance:
replicas: 1
disable-ha: false
buckets: 1
threads-per-controller: 2
kube-api-qps: 5.0
kube-api-burst: 10
Chain
Chain section allows user to customize the Tekton Chain features. This allows user to customize the values in configmaps
Refer to properties section in TektonChain for available options
Example:
chain:
disabled: false # - `disabled` : if the value set as `true`, chains feature will be disabled (default: `false`)
targetNamespace: tekton-pipelines
controllerEnvs:
- name: MONGO_SERVER_URL # This is the only field supported at the moment which is optional and when added by user, it is added as env to Chains controller
value: #value # This can be provided same as env field of container
artifacts.taskrun.format: in-toto
artifacts.taskrun.storage: tekton,oci (comma separated values)
artifacts.taskrun.signer: x509
artifacts.oci.storage: oci (comma separated values)
artifacts.oci.format: simplesigning
artifacts.oci.signer: x509
artifacts.pipelinerun.format: in-toto
artifacts.pipelinerun.storage: tekton,oci (comma separated values)
artifacts.pipelinerun.signer: x509
storage.gcs.bucket: #value
storage.oci.repository: #value
storage.oci.repository.insecure: #value (boolean - true/false)
storage.docdb.url: #value
storage.grafeas.projectid: #value
storage.grafeas.noteid: #value
storage.grafeas.notehint: #value
builder.id: #value
signers.x509.fulcio.enabled: #value (boolean - true/false)
signers.x509.fulcio.address: #value
signers.x509.fulcio.issuer: #value
signers.x509.fulcio.provider: #value
signers.x509.identity.token.file: #value
signers.x509.tuf.mirror.url: #value
signers.kms.kmsref: #value
signers.kms.kmsref.auth.address: #value
signers.kms.kmsref.auth.token: #value
signers.kms.kmsref.auth.oidc.path: #value
signers.kms.kmsref.auth.oidc.role: #value
signers.kms.kmsref.auth.spire.sock: #value
signers.kms.kmsref.auth.spire.audience: #value
transparency.enabled: #value (boolean - true/false)
transparency.url: #value
Pruner
Pruner provides auto clean up feature for the Tekton pipelinerun
and taskrun
resources. In the background pruner container runs tkn
command.
Example:
pruner:
disabled: false
schedule: "0 8 * * *"
resources:
- taskrun
- pipelinerun
keep: 3
# keep-since: 1440
# NOTE: you can use either "keep" or "keep-since", not both
prune-per-resource: true
disabled
: if the value set astrue
, pruner feature will be disabled (default:false
)schedule
: how often to run the pruner job. User can understand the schedule syntax here.resources
: supported resources for auto prune aretaskrun
andpipelinerun
keep
: maximum number of resources to keep while deleting or removing resourceskeep-since
: retain the resources younger than the specified value in minutesprune-per-resource
: if the value set astrue
(default valuefalse
), thekeep
applied to each resource. The resources(pipeline
and/ortask
) taken dynamically from that namespace and applied.
example: in a namespacens-1
I have twopipeline
, namedpipeline-1
andpipeline-2
, the out come would be:tkn pipelinerun delete --pipeline=my-pipeline-1 --keep=3 --namespace=ns-1
,tkn pipelinerun delete --pipeline=my-pipeline-2 --keep=3 --namespace=ns-1
. the same way works fortask
too.
We do not see any benefit by enablingprune-per-resource=true
, when you usekeep-since
. Askeep-since
is limiting the resources by time(irrespective of resource count), there is no change on the outcome.
Note:
if
disabled: false
andschedule:
with empty value, global pruner job will be disabled. however, if there is a prune schedule (operator.tekton.dev/prune.schedule
) annotation present with a value in a namespace. a namespace wide pruner jobs will be created.
Pruner Namespace annotations
By default pruner job will be created from the global pruner config (spec.pruner
), though user can customize a pruner config to a specific namespace with the following annotations. If some of the annotations are not present or has invalid value, for that value, falls back to global value or skipped the namespace.
operator.tekton.dev/prune.skip
- pruner job will be skipped to a namespace, if the value set astrue
operator.tekton.dev/prune.schedule
- pruner job will be created on a specific scheduleoperator.tekton.dev/prune.keep
- maximum number of resources will be keptoperator.tekton.dev/prune.keep-since
- retain the resources younger than the specified value in minutesoperator.tekton.dev/prune.prune-per-resource
- thekeep
orkeep-since
applied to each resourceoperator.tekton.dev/prune.resources
- can betaskrun
and/orpipelinerun
, both value can be specified with comma separated. example:taskrun,pipelinerun
operator.tekton.dev/prune.strategy
- allowed values: eitherkeep
orkeep-since
Note:
if a global value is not present the following values will be consider as default value
resources: pipelinerun
keep: 100
Addon
TektonAddon install some resources along with Tekton Pipelines on the cluster. This provides few ClusterTasks, PipelineTemplates.
This section allows to customize installation of those resources through params. You can read more about the supported params here.
Example:
addon:
params:
- name: "clusterTask"
value: "true"
- name: "pipelineTemplates"
value: "true"
NOTE: TektonAddon is currently available for OpenShift Platform only. Enabling this for Kubernetes platform is in roadmap of Operator.
Hub
This is to enable/disable showing hub resources in pipeline builder of devconsole(OpenShift UI). By default, the field is
not there in the config object. If you want to disable the integration, you can add the param like below in config with value false
.
The possible values are true
and false
.
Example:
hub:
params:
- name: enable-devconsole-integration
value: "false"
Dashboard
Dashboard provides configuration options for the Tekton Dashboard if the specified profile value includes the Dashboard component. (E.g. all
on Kubernetes)
Example:
dashboard:
readonly: true
readonly
: If set to true, install the Dashboard in read-only mode
This is an Optional
section.
Resolvers
As part of TektonPipelines, resolvers are installed which are by default enabled. User can disable them through TektonConfig.
apiVersion: operator.tekton.dev/v1alpha1
kind: TektonConfig
metadata:
name: config
spec:
pipeline:
enable-bundles-resolver: true
enable-cluster-resolver: true
enable-git-resolver: true
enable-hub-resolver: true
User can also provide resolver specific configurations through TektonConfig. The Default configurations are not added by default in TektonConfig. To override default configurations, user can provide the configurations as below. You can find the default configurations here
apiVersion: operator.tekton.dev/v1alpha1
kind: TektonConfig
metadata:
name: config
spec:
pipeline:
bundles-resolver-config:
default-service-account: pipelines
cluster-resolver-config:
default-namespace: cluster-resolver-test
enable-bundles-resolver: true
enable-cluster-resolver: true
enable-git-resolver: true
enable-hub-resolver: true
git-resolver-config:
server-url: localhost.com
hub-resolver-config:
default-tekton-hub-catalog: tekton
tekton-hub-api: "https://my-custom-tekton-hub.example.com"
artifact-hub-api: "https://my-custom-artifact-hub.example.com"
Under the hub-resolver-config
, the tekton-hub-api
and artifact-hub-api
will be passed as environment variable into tekton-pipelines-remote-resolvers
controller.
In the deployment the environment name will be converted as follows,
tekton-hub-api
=>TEKTON_HUB_API
artifact-hub-api
=>ARTIFACT_HUB_API
OpenShiftPipelinesAsCode
The PipelinesAsCode section allows you to customize the Pipelines as Code features. When you change the TektonConfig CR, the Operator automatically applies the settings to custom resources and configmaps in your installation.
Some of the fields have default values, so operator will add them if the user hasn’t passed in CR. Other fields which don’t have default values unless the user specifies them. User can find those here.
Example:
platforms:
openshift:
pipelinesAsCode:
enable: true
settings:
application-name: Pipelines as Code CI
auto-configure-new-github-repo: "false"
bitbucket-cloud-check-source-ip: "true"
custom-console-name: ""
custom-console-url: ""
custom-console-url-pr-details: ""
custom-console-url-pr-tasklog: ""
error-detection-from-container-logs: "false"
error-detection-max-number-of-lines: "50"
error-detection-simple-regexp: ^(?P<filename>[^:]*):(?P<line>[0-9]+):(?P<column>[0-9]+):([
]*)?(?P<error>.*)
error-log-snippet: "true"
hub-catalog-name: tekton
hub-url: https://api.hub.tekton.dev/v1
remote-tasks: "true"
secret-auto-create: "true"
secret-github-app-token-scoped: "true"
NOTE: OpenShiftPipelinesAsCode is currently available for the OpenShift Platform only.
Additional fields as options
There is a filed called options
available in all the components.
NOTE: There is a possibility to have two different values for a field.
An example: with a pre-defined field you can set value and the same filed may be defined underoptions
as well. In that case value fromoptions
will be final.
options
filed is defined as follows,
options:
disabled: false
configMaps:
config-leader-election: # name of the configMap
data:
lease-duration: "90s"
pipeline-config-logging: # creates new configMap under targetNamespace
metadata:
labels:
my-custom-logger: "true"
annotations:
logger-type: "uber-zap"
data:
loglevel.controller: "info"
loglevel.webhook: "info"
zap-logger-config: |
{
"level": "debug",
"development": false,
"sampling": {
"initial": 100,
"thereafter": 50
},
"outputPaths": ["stdout"],
"errorOutputPaths": ["stderr"],
"encoding": "json",
"encoderConfig": {
"timeKey": "ts",
"levelKey": "severity",
"nameKey": "logger",
"callerKey": "caller",
"messageKey": "message",
"stacktraceKey": "stacktrace",
"lineEnding": "",
"levelEncoder": "",
"timeEncoder": "iso8601",
"durationEncoder": "",
"callerEncoder": ""
}
}
deployments:
tekton-pipelines-controller:
metadata:
labels:
custom-label: "foo"
annotations:
custom-annotation: "foo"
spec:
replicas: 2
template:
spec:
containers:
- name: tekton-pipelines-controller
env:
- name: CONFIG_LOGGING_NAME
value: pipeline-config-logging
statefulSets:
web:
metadata:
labels:
custom-label: foo
annotations:
custom-annotation: foo
spec:
replicas: 3
template:
spec:
containers:
- name: nginx
env:
- name: NGINX_MODE
value: production
disabled
- disables the additionaloptions
support, ifdisabled
set astrue
. default:false
ConfigMaps
Supports to update existing configMap also supports to create new configMap.
The following fields are supported in configMap
metadata
labels
- supports add and updateannotations
- supports add and update
data
- supports add and update
Deployments
Supports to update the existing deployments. But not supported to create new deployment.
The following fields are supported in deployment
metadata
labels
- supports add and updateannotations
- supports add and update
spec
replicas
- updates deployment replicas counttemplate
spec
affinity
- replaces the existing Affinity with this, if not emptypriorityClassName
- replaces the existing PriorityClassName with this, if not emptynodeSelector
- replaces the existing NodeSelector with this, if not emptytolerations
- replaces the existing tolerations with this, if not emptytopologySpreadConstraints
- replaces the existing TopologySpreadConstraints with this, if not emptyvolumes
- adds and updates volumesinitContainers
- updates init-containersresources
- replaces the resources requirements with this, if not emptyenvs
- adds and updates environmentsvolumeMounts
- adds and updates VolumeMountsargs
- appends given args with existing arguments. NOTE: THIS OPERATION DO NOT REPLACE EXISTING ARGS
containers
- updates containersresources
- replaces the resources requirements with this, if not emptyenvs
- adds and updates environmentsvolumeMounts
- adds and updates VolumeMountsargs
- appends given args with existing arguments. NOTE: THIS OPERATION DO NOT REPLACE EXISTING ARGS
StatefulSets
Supports to update the existing StatefulSet. But not supported to create new StatefulSet.
The following fields are supported in StatefulSet
metadata
labels
- supports add and updateannotations
- supports add and update
spec
replicas
- updates statefulSets replicas countserviceName
- updates service namepodManagementPolicy
- updates pod management policyvolumeClaimTemplates
- updates volume claim templatestemplate
spec
affinity
- replaces the existing Affinity with this, if not emptypriorityClassName
- replaces the existing PriorityClassName with this, if not emptynodeSelector
- replaces the existing NodeSelector with this, if not emptytolerations
- replaces the existing tolerations with this, if not emptytopologySpreadConstraints
- replaces the existing TopologySpreadConstraints with this, if not emptyvolumes
- adds and updates volumesinitContainers
- updates init-containersresources
- replaces the resources requirements with this, if not emptyenvs
- adds and updates environmentsvolumeMounts
- adds and updates VolumeMountsargs
- appends given args with existing arguments. NOTE: THIS OPERATION DO NOT REPLACE EXISTING ARGS
containers
- updates containersresources
- replaces the resources requirements with this, if not emptyenvs
- adds and updates environmentsvolumeMounts
- adds and updates VolumeMountsargs
- appends given args with existing arguments. NOTE: THIS OPERATION DO NOT REPLACE EXISTING ARGS
NOTE: If a Deployment or StatefulSet has a Horizontal Pod Autoscaling (HPA) and is in active state, Operator will not control the replicas to that resource. However if status.desiredReplicas
and spec.minReplicas
not present in HPA, operator takes the control. Also if HPA disabled, operator takes control. Even though the operator takes the control, the replicas value will be adjusted to the hpa’s scaling range.
Feedback
Was this page helpful?