OLM deployment resources are templated so that they can be easily configured for different deployment environments.
kubectl create -f deploy/upstream/manifests/latest/
oc create namespace operator-lifecycle-manager
oc create -f deploy/okd/manifests/latest/
OLM should be installed via openshift-ansible. The manifests in this repo is periodically synced with openshift-ansible and should only be used for testing releases.
This command starts minikube, builds the OLM containers locally with the minikube-provided docker, and uses the local configuration in local-values.yaml to build localized deployment resources for OLM.
make run-local
You can verify that the OLM components have been successfully deployed by running kubectl -n local get deployments
This command starts minishift, builds the OLM containers locally with the minishift-provided docker, and uses the local configuration in local-values-shift.yaml to build localized deployment resources for OLM.
make run-local-shift
You can verify that the OLM components have been successfully deployed by running kubectl -n local get deployments
Deployments of OLM can be stamped out with different configurations by writing a values.yaml
file and running commands to generate resources.
Here's an example values.yaml
# sets the apiversion to use for rbac-resources. Change to `authorization.openshift.io` for openshift
rbacApiVersion: rbac.authorization.k8s.io
# namespace is the namespace the operators will _run_
namespace: local
# watchedNamespaces is a comma-separated list of namespaces the operators will _watch_ for OLM resources.
# Omit to enable OLM in all namespaces
watchedNamespaces: local
# catalog_namespace is the namespace where the catalog operator will look for global catalogs.
# entries in global catalogs can be resolved in any watched namespace
catalog_namespace: local
# OLM operator run configuration
alm:
# OLM operator doesn't do any leader election (yet), set to 1
replicaCount: 1
# The image to run. If not building a local image, use sha256 image references
image:
ref: quay.io/coreos/olm:local
pullPolicy: IfNotPresent
service:
# port for readiness/liveness probes
internalPort: 8080
# catalog operator run configuration
catalog:
# Catalog operator doesn't do any leader election (yet), set to 1
replicaCount: 1
# The image to run. If not building a local image, use sha256 image references
image:
ref: quay.io/coreos/catalog:local
pullPolicy: IfNotPresent
service:
# port for readiness/liveness probes
internalPort: 8080
To configure a release of OLM for installation in a cluster:
- Create a
my-values.yaml
like the example above with the desired configuration or choose an existing one from this repository. The latest production values can be found in deploy/tectonic-alm-operator/values.yaml. - Generate deployment files from the templates and the
my-values.yaml
usingpackage-release.sh
# first arg must be a semver-compatible version string # second arg is the output directory # third arg is the values.yaml file ./scripts/package-release.sh 1.0.0-myolm ./my-olm-deployment my-values.yaml
- Deploy to kubernetes:
kubectl apply -f ./my-olm-deployment/templates/
The above steps are automated for official releases with make ver=0.3.0 release
, which will output new versions of manifests in deploy/tectonic-alm-operator/manifests/$(ver)
.
Cloud Services can be installed from the catalog by subscribing to a channel in the corresponding package.
If using one of the local
run options, this will subscribe to etcd
, vault
, and prometheus
operators. Subscribing to a service that doesn't exist yet will install the operator and related CRDs in the namespace.
apiVersion: operators.coreos.com/v1alpha1
kind: Subscription
metadata:
name: etcd
namespace: local
spec:
channel: alpha
name: etcd
source: rh-operators
---
apiVersion: operators.coreos.com/v1alpha1
kind: Subscription
metadata:
name: prometheus
namespace: local
spec:
channel: alpha
name: prometheus
source: rh-operators