diff --git a/Makefile b/Makefile index b2357975a45..430584e9cbd 100644 --- a/Makefile +++ b/Makefile @@ -138,9 +138,7 @@ endef .PHONY: manifests manifests: controller-gen ## Generate WebhookConfiguration, ClusterRole and CustomResourceDefinition objects. -# TODO: enable below when we do webhook -# $(CONTROLLER_GEN) rbac:roleName=controller-manager-role crd webhook paths="./..." output:crd:artifacts:config=config/crd/bases - $(CONTROLLER_GEN) rbac:roleName=controller-manager-role crd:ignoreUnexportedFields=true paths="./..." output:crd:artifacts:config=config/crd/bases + $(CONTROLLER_GEN) rbac:roleName=controller-manager-role crd:ignoreUnexportedFields=true webhook paths="./..." output:crd:artifacts:config=config/crd/bases $(call fetch-external-crds,github.com/openshift/api,route/v1) $(call fetch-external-crds,github.com/openshift/api,user/v1) diff --git a/PROJECT b/PROJECT index cbb40d6a5b2..4bfc6c97792 100644 --- a/PROJECT +++ b/PROJECT @@ -21,6 +21,9 @@ resources: kind: DSCInitialization path: github.com/opendatahub-io/opendatahub-operator/v2/apis/dscinitialization/v1 version: v1 + webhooks: + validation: true + webhookVersion: v1 - api: crdVersion: v1 namespaced: false @@ -30,4 +33,7 @@ resources: kind: DataScienceCluster path: github.com/opendatahub-io/opendatahub-operator/v2/apis/datasciencecluster/v1 version: v1 + webhooks: + validation: true + webhookVersion: v1 version: "3" diff --git a/bundle/manifests/datasciencecluster.opendatahub.io_datascienceclusters.yaml b/bundle/manifests/datasciencecluster.opendatahub.io_datascienceclusters.yaml index 457c39ddf59..45c687adb48 100644 --- a/bundle/manifests/datasciencecluster.opendatahub.io_datascienceclusters.yaml +++ b/bundle/manifests/datasciencecluster.opendatahub.io_datascienceclusters.yaml @@ -62,12 +62,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -105,12 +105,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -149,12 +149,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -195,12 +195,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -298,12 +298,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -342,12 +342,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -385,12 +385,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -428,12 +428,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array @@ -471,12 +471,12 @@ spec: description: 'sourcePath is the subpath within contextDir where kustomize builds start. Examples include any sub-folder or path: `base`, `overlays/dev`, - `default`, `odh` etc' + `default`, `odh` etc.' type: string uri: default: "" description: uri is the URI point to a git repo - with tag/branch. e.g https://github.com/org/repo/tarball/ + with tag/branch. e.g. https://github.com/org/repo/tarball/ type: string type: object type: array diff --git a/bundle/manifests/dscinitialization.opendatahub.io_dscinitializations.yaml b/bundle/manifests/dscinitialization.opendatahub.io_dscinitializations.yaml index 8ae3397ba85..6b9514bb1e5 100644 --- a/bundle/manifests/dscinitialization.opendatahub.io_dscinitializations.yaml +++ b/bundle/manifests/dscinitialization.opendatahub.io_dscinitializations.yaml @@ -131,6 +131,7 @@ spec: field. properties: customCABundle: + default: "" description: A custom CA bundle that will be available for all components in the Data Science Cluster(DSC). This bundle will be stored in odh-trusted-ca-bundle ConfigMap .data.odh-ca-bundle.crt . @@ -146,6 +147,7 @@ spec: pattern: ^(Managed|Unmanaged|Force|Removed)$ type: string required: + - customCABundle - managementState type: object required: diff --git a/bundle/manifests/redhat-ods-operator-webhook-service_v1_service.yaml b/bundle/manifests/redhat-ods-operator-webhook-service_v1_service.yaml new file mode 100644 index 00000000000..218c45fefcc --- /dev/null +++ b/bundle/manifests/redhat-ods-operator-webhook-service_v1_service.yaml @@ -0,0 +1,24 @@ +apiVersion: v1 +kind: Service +metadata: + annotations: + service.beta.openshift.io/inject-cabundle: "true" + service.beta.openshift.io/serving-cert-secret-name: redhat-ods-operator-controller-webhook-cert + creationTimestamp: null + labels: + app.kubernetes.io/component: webhook + app.kubernetes.io/created-by: rhods-operator + app.kubernetes.io/instance: webhook-service + app.kubernetes.io/managed-by: kustomize + app.kubernetes.io/name: service + app.kubernetes.io/part-of: rhods-operator + name: redhat-ods-operator-webhook-service +spec: + ports: + - port: 443 + protocol: TCP + targetPort: 9443 + selector: + control-plane: controller-manager +status: + loadBalancer: {} diff --git a/bundle/manifests/rhods-operator.clusterserviceversion.yaml b/bundle/manifests/rhods-operator.clusterserviceversion.yaml index 9417d9431be..4817d94ac40 100644 --- a/bundle/manifests/rhods-operator.clusterserviceversion.yaml +++ b/bundle/manifests/rhods-operator.clusterserviceversion.yaml @@ -106,9 +106,9 @@ metadata: capabilities: Full Lifecycle categories: AI/Machine Learning, Big Data certified: "False" - containerImage: quay.io/opendatahub/opendatahub-operator:v2.4.0 + containerImage: quay.io/opendatahub/opendatahub-operator:v2.7.0 createdAt: "2023-8-23T00:00:00Z" - olm.skipRange: '>=1.0.0 <2.0.0' + olm.skipRange: '>=1.0.0 <2.7.0' operatorframework.io/initialization-resource: |- { "apiVersion": "datasciencecluster.opendatahub.io/v1", @@ -157,7 +157,7 @@ metadata: }, "workbenches": { "managementState": "Managed" - } + }, } } } @@ -165,7 +165,7 @@ metadata: operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]' operators.operatorframework.io/project_layout: go.kubebuilder.io/v3 repository: https://github.com/red-hat-data-services/rhods-operator - name: rhods-operator.v2.4.0 + name: rhods-operator.v2.8.0 namespace: placeholder spec: apiservicedefinitions: {} @@ -218,28 +218,7 @@ spec: - kind: FeatureTracker name: featuretrackers.features.opendatahub.io version: v1 - description: |- - Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects. - - When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster. - - Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration. - - But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements. - - Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference. - - ### Components - * Dashboard - * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, VScode) - * Ability to add Custom Images - * Ability to leverage accelerators (such as NVIDIA GPU) - * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines) - * Model Serving using ModelMesh and Kserve. - * Ability to use other runtimes for serving - * Model Monitoring - * Distributed workloads (KubeRay, CodeFlare, Kueue) - * XAI explanations of predictive models (TrustyAI) + description: This will be replaced by Kustomize displayName: Red Hat OpenShift AI icon: - base64data: <?xml version="1.0" encoding="utf-8"?>
<!-- Generator: Adobe Illustrator 27.4.1, SVG Export Plug-In . SVG Version: 6.00 Build 0)  -->
<svg version="1.1" id="Logos" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" x="0px" y="0px"
	 viewBox="0 0 835.9 244" style="enable-background:new 0 0 835.9 244;" xml:space="preserve">
<style type="text/css">
	.st0{fill:#FFFFFF;}
	.st1{fill:#EE0000;}
</style>
<g>
	<g>
		<path d="M301.3,183.1c0,5.2-1,10.1-3,14.7c-2,4.6-4.7,8.6-8.1,12c-3.4,3.4-7.4,6-12.1,8c-4.6,1.9-9.6,2.9-14.9,2.9
			c-5.3,0-10.2-1-14.9-2.9c-4.6-1.9-8.7-4.6-12-8c-3.4-3.4-6-7.3-8-12c-2-4.6-3-9.5-3-14.7c0-5.2,1-10.1,3-14.7c2-4.6,4.6-8.6,8-12
			c3.4-3.4,7.4-6,12-8c4.6-1.9,9.6-2.9,14.9-2.9c5.3,0,10.2,1,14.9,2.9c4.6,1.9,8.7,4.6,12.1,8c3.4,3.4,6.1,7.4,8.1,12
			C300.4,173,301.3,177.9,301.3,183.1z M290.2,183.1c0-3.9-0.7-7.5-2.1-10.9c-1.4-3.4-3.3-6.3-5.7-8.7c-2.4-2.5-5.2-4.4-8.5-5.8
			c-3.3-1.4-6.8-2.1-10.6-2.1s-7.2,0.7-10.5,2.1c-3.3,1.4-6.1,3.3-8.5,5.8c-2.4,2.5-4.3,5.4-5.7,8.7c-1.4,3.4-2.1,7-2.1,10.9
			c0,3.9,0.7,7.5,2.1,10.9c1.4,3.4,3.3,6.3,5.7,8.7c2.4,2.4,5.2,4.4,8.5,5.8c3.3,1.4,6.8,2.1,10.5,2.1s7.3-0.7,10.6-2.1
			c3.3-1.4,6.1-3.3,8.5-5.8c2.4-2.4,4.3-5.3,5.7-8.7C289.5,190.6,290.2,187,290.2,183.1z"/>
		<path d="M311.6,241.1v-74.5h10.3v5c2.2-1.9,4.7-3.3,7.5-4.3c2.8-1,5.7-1.5,8.7-1.5c3.7,0,7.2,0.7,10.5,2.1
			c3.3,1.4,6.1,3.4,8.5,5.8c2.4,2.5,4.3,5.4,5.7,8.7c1.4,3.3,2.1,6.9,2.1,10.6c0,3.8-0.7,7.4-2.1,10.7c-1.4,3.3-3.3,6.2-5.7,8.7
			c-2.4,2.5-5.3,4.4-8.6,5.8c-3.3,1.4-6.9,2.1-10.7,2.1c-3,0-5.8-0.5-8.5-1.4c-2.7-0.9-5.1-2.2-7.3-3.8v25.9H311.6z M336.7,174.8
			c-3.1,0-5.8,0.6-8.3,1.7c-2.5,1.1-4.6,2.6-6.3,4.6v24.1c1.7,1.9,3.8,3.4,6.3,4.5c2.6,1.1,5.3,1.7,8.3,1.7c5.1,0,9.4-1.8,12.8-5.3
			c3.4-3.5,5.1-7.8,5.1-12.9c0-5.2-1.8-9.6-5.3-13.1C345.9,176.6,341.7,174.8,336.7,174.8z"/>
		<path d="M372.4,193c0-3.7,0.7-7.3,2-10.6c1.4-3.3,3.2-6.2,5.6-8.7c2.4-2.5,5.2-4.4,8.4-5.8c3.2-1.4,6.7-2.1,10.5-2.1
			c3.6,0,7,0.7,10.1,2.1c3.2,1.4,5.9,3.4,8.1,5.8c2.3,2.5,4,5.4,5.4,8.8c1.3,3.4,2,7,2,10.9v3h-41.8c0.7,4.4,2.7,8,6,10.9
			c3.3,2.9,7.3,4.3,11.9,4.3c2.6,0,5-0.4,7.4-1.2c2.4-0.8,4.4-2,6-3.4l6.7,6.6c-3.1,2.4-6.3,4.2-9.6,5.3c-3.3,1.1-6.9,1.7-10.9,1.7
			c-3.9,0-7.5-0.7-10.9-2.1c-3.4-1.4-6.3-3.3-8.8-5.8c-2.5-2.4-4.5-5.3-5.9-8.7C373.1,200.5,372.4,196.9,372.4,193z M398.7,174.5
			c-4,0-7.5,1.3-10.4,4c-2.9,2.6-4.8,6-5.5,10.2h31.4c-0.7-4-2.5-7.4-5.4-10.1C405.9,175.9,402.5,174.5,398.7,174.5z"/>
		<path d="M434.3,219.5v-52.9h10.4v5.3c2.1-2.1,4.5-3.7,7.1-4.7c2.7-1.1,5.6-1.6,8.8-1.6c6,0,11,1.9,14.8,5.8
			c3.8,3.9,5.8,8.8,5.8,14.9v33.3h-10.3V188c0-4.1-1.2-7.3-3.5-9.8c-2.4-2.4-5.6-3.6-9.7-3.6c-2.8,0-5.3,0.6-7.5,1.8
			c-2.2,1.2-4.1,2.9-5.5,5.1v38.1H434.3z"/>
		<path d="M488,207.5l6.7-7.7c3.9,3.8,7.9,6.7,12,8.6c4.1,1.9,8.5,2.9,13.1,2.9c5.3,0,9.5-1.1,12.8-3.4c3.3-2.3,4.9-5.2,4.9-8.7
			c0-3.2-1.1-5.7-3.3-7.4c-2.2-1.8-6-3.1-11.4-4l-12.2-2c-6.7-1.1-11.7-3.3-15-6.4c-3.3-3.2-4.9-7.3-4.9-12.5
			c0-6.2,2.4-11.3,7.3-15.1c4.9-3.8,11.4-5.8,19.5-5.8c5.1,0,10.3,0.8,15.4,2.5c5.1,1.7,9.8,4.1,13.9,7.3l-6,8.3
			c-4-3-7.9-5.2-11.9-6.7c-4-1.5-8-2.2-11.9-2.2c-4.7,0-8.5,1-11.4,3c-2.9,2-4.4,4.6-4.4,7.8c0,2.9,1,5.2,3,6.8
			c2,1.6,5.3,2.8,10,3.5l11.8,1.9c7.7,1.2,13.3,3.5,17,6.8c3.6,3.3,5.4,7.7,5.4,13.4c0,3.3-0.7,6.4-2.1,9.1
			c-1.4,2.7-3.3,5.1-5.9,7.1c-2.5,2-5.6,3.5-9.2,4.6c-3.6,1.1-7.6,1.6-11.9,1.6c-5.8,0-11.4-1.1-16.8-3.4
			C497,214.9,492.2,211.6,488,207.5z"/>
		<path d="M567.4,144.5v75.1H557v-72.8L567.4,144.5z M557,219.5v-52.9h10.4v5.3c2.1-2.1,4.5-3.7,7.1-4.7c2.7-1.1,5.6-1.6,8.8-1.6
			c6,0,11,1.9,14.8,5.8c3.8,3.9,5.8,8.8,5.8,14.9v33.3h-10.3V188c0-4.1-1.2-7.3-3.5-9.8c-2.4-2.4-5.6-3.6-9.7-3.6
			c-2.8,0-5.3,0.6-7.5,1.8c-2.2,1.2-4.1,2.9-5.5,5.1v38.1H557z"/>
		<path d="M621.1,158.1c-1.7,0-3.2-0.6-4.5-1.9c-1.3-1.3-1.9-2.8-1.9-4.5c0-1.7,0.6-3.2,1.9-4.5c1.3-1.3,2.8-1.9,4.5-1.9
			c1.7,0,3.2,0.6,4.5,1.9c1.2,1.3,1.9,2.8,1.9,4.5c0,1.7-0.6,3.2-1.9,4.5C624.4,157.4,622.9,158.1,621.1,158.1z M626.3,166.6v52.9
			h-10.4v-52.9H626.3z"/>
		<path d="M634.1,166.6h12.3v-8c0-5.3,1.5-9.5,4.6-12.5c3-3,7.5-4.5,13.4-4.5c1.3,0,2.6,0.1,3.9,0.3c1.3,0.2,2.5,0.4,3.6,0.7v9
			c-1.2-0.3-2.3-0.6-3.2-0.7c-1-0.1-2.1-0.2-3.3-0.2c-2.9,0-5.1,0.7-6.5,2c-1.4,1.3-2.1,3.4-2.1,6.2v7.8h15.2v8.7h-15.2v44.2h-10.3
			v-44.2h-12.3V166.6z"/>
		<path d="M687.7,206.4v-31.1h-11.2v-8.7h11.2v-13.5l10.3-2.5v16h15.6v8.7H698V204c0,2.7,0.6,4.6,1.8,5.7c1.2,1.1,3.2,1.7,6,1.7
			c1.5,0,2.8-0.1,4-0.3c1.1-0.2,2.3-0.5,3.6-1v8.7c-1.5,0.5-3.1,0.9-4.9,1.1c-1.8,0.3-3.5,0.4-5,0.4c-5.1,0-9-1.2-11.8-3.6
			S687.7,211,687.7,206.4z"/>
		<path d="M737,219.5l30.2-72.8h12.8l29.7,72.8h-11.9l-8.4-21.3h-32.6l-8.5,21.3H737z M760.5,189.2h25.4l-12.7-31.9L760.5,189.2z"/>
		<path d="M817.2,219.5v-72.8h10.9v72.8H817.2z"/>
	</g>
	<g>
		<g>
			<path class="st1" d="M129,85c12.5,0,30.6-2.6,30.6-17.5c0-1.2,0-2.3-0.3-3.4l-7.4-32.4c-1.7-7.1-3.2-10.3-15.7-16.6
				C126.4,10.2,105.3,2,99,2c-5.8,0-7.5,7.5-14.4,7.5c-6.7,0-11.6-5.6-17.9-5.6c-6,0-9.9,4.1-12.9,12.5c0,0-8.4,23.7-9.5,27.2
				C44,44.3,44,45,44,45.5C44,54.8,80.3,85,129,85 M161.5,73.6c1.7,8.2,1.7,9.1,1.7,10.1c0,14-15.7,21.8-36.4,21.8
				C80,105.5,39.1,78.1,39.1,60c0-2.8,0.6-5.4,1.5-7.3C23.8,53.5,2,56.5,2,75.7C2,107.2,76.6,146,135.7,146
				c45.3,0,56.7-20.5,56.7-36.6C192.3,96.6,181.4,82.2,161.5,73.6"/>
			<path d="M161.5,73.6c1.7,8.2,1.7,9.1,1.7,10.1c0,14-15.7,21.8-36.4,21.8C80,105.5,39.1,78.1,39.1,60c0-2.8,0.6-5.4,1.5-7.3
				l3.7-9.1C44,44.3,44,45,44,45.5C44,54.8,80.3,85,129,85c12.5,0,30.6-2.6,30.6-17.5c0-1.2,0-2.3-0.3-3.4L161.5,73.6z"/>
		</g>
		<path d="M581.2,94.3c0,11.9,7.2,17.7,20.2,17.7c3.2,0,8.6-0.7,11.9-1.7V96.5c-2.8,0.8-4.9,1.2-7.7,1.2c-5.4,0-7.4-1.7-7.4-6.7
			V69.8h15.6V55.6h-15.6v-18l-17,3.7v14.3H570v14.2h11.3V94.3z M528.3,94.6c0-3.7,3.7-5.5,9.3-5.5c3.7,0,7,0.5,10.1,1.3v7.2
			c-3.2,1.8-6.8,2.6-10.6,2.6C531.6,100.2,528.3,98.1,528.3,94.6 M533.5,112.2c6,0,10.8-1.3,15.4-4.3v3.4h16.8V75.6
			c0-13.6-9.1-21-24.4-21c-8.5,0-16.9,2-26,6.1l6.1,12.5c6.5-2.7,12-4.4,16.8-4.4c7,0,10.6,2.7,10.6,8.3v2.7
			c-4-1.1-8.2-1.6-12.6-1.6c-14.3,0-22.9,6-22.9,16.7C513.3,104.7,521.1,112.2,533.5,112.2 M441.1,111.2h18.1V82.4h30.3v28.8h18.1
			V37.6h-18.1v28.3h-30.3V37.6h-18.1V111.2z M372.1,83.4c0-8,6.3-14.1,14.6-14.1c4.6,0,8.8,1.6,11.8,4.3V93c-3,2.9-7,4.4-11.8,4.4
			C378.5,97.5,372.1,91.4,372.1,83.4 M398.7,111.2h16.8V33.9l-17,3.7v20.9c-4.2-2.4-9-3.7-14.2-3.7c-16.2,0-28.9,12.5-28.9,28.5
			c0,16,12.5,28.6,28.4,28.6c5.5,0,10.6-1.7,14.9-4.8V111.2z M321.5,68.5c5.4,0,9.9,3.5,11.7,8.8H310
			C311.7,71.8,315.9,68.5,321.5,68.5 M292.8,83.5c0,16.2,13.3,28.8,30.3,28.8c9.4,0,16.2-2.5,23.2-8.4l-11.3-10
			c-2.6,2.7-6.5,4.2-11.1,4.2c-6.3,0-11.5-3.5-13.7-8.8h39.6V85c0-17.7-11.9-30.4-28.1-30.4C305.6,54.7,292.8,67.3,292.8,83.5
			 M263.5,53.1c6,0,9.4,3.8,9.4,8.3s-3.4,8.3-9.4,8.3h-17.9V53.1H263.5z M227.5,111.2h18.1V84.4h13.8l13.9,26.8h20.2l-16.2-29.4
			c8.7-3.8,13.9-11.7,13.9-20.7c0-13.3-10.4-23.5-26-23.5h-37.7V111.2z"/>
	</g>
</g>
</svg>
 @@ -1744,18 +1723,6 @@ spec: - patch - update - watch - - apiGroups: - - authentication.k8s.io - resources: - - tokenreviews - verbs: - - create - - apiGroups: - - authorization.k8s.io - resources: - - subjectaccessreviews - verbs: - - create serviceAccountName: redhat-ods-operator-controller-manager deployments: - label: @@ -1790,6 +1757,10 @@ spec: initialDelaySeconds: 15 periodSeconds: 20 name: manager + ports: + - containerPort: 9443 + name: webhook-server + protocol: TCP readinessProbe: httpGet: path: /readyz @@ -1803,44 +1774,24 @@ spec: requests: cpu: 500m memory: 256Mi + securityContext: + allowPrivilegeEscalation: false + capabilities: + drop: + - ALL + volumeMounts: + - mountPath: /tmp/k8s-webhook-server/serving-certs + name: cert + readOnly: true securityContext: runAsNonRoot: true serviceAccountName: redhat-ods-operator-controller-manager terminationGracePeriodSeconds: 10 - permissions: - - rules: - - apiGroups: - - "" - resources: - - configmaps - verbs: - - get - - list - - watch - - create - - update - - patch - - delete - - apiGroups: - - coordination.k8s.io - resources: - - leases - verbs: - - get - - list - - watch - - create - - update - - patch - - delete - - apiGroups: - - "" - resources: - - events - verbs: - - create - - patch - serviceAccountName: redhat-ods-operator-controller-manager + volumes: + - name: cert + secret: + defaultMode: 420 + secretName: redhat-ods-operator-controller-webhook-cert strategy: deployment installModes: - supported: false @@ -1861,11 +1812,34 @@ spec: - training - kserve - distributed-workloads - - trustyai links: - name: Red Hat OpenShift AI url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai minKubeVersion: 1.22.0 provider: name: Red Hat - version: 2.4.0 + replaces: rhods-operator.v2.6.0 + version: 2.8.0 + webhookdefinitions: + - admissionReviewVersions: + - v1 + containerPort: 443 + deploymentName: redhat-ods-operator-controller-manager + failurePolicy: Fail + generateName: operator.opendatahub.io + rules: + - apiGroups: + - datasciencecluster.opendatahub.io + - dscinitialization.opendatahub.io + apiVersions: + - v1 + operations: + - CREATE + - UPDATE + resources: + - datascienceclusters + - dscinitializations + sideEffects: None + targetPort: 9443 + type: ValidatingAdmissionWebhook + webhookPath: /validate-opendatahub-io-v1 diff --git a/config/crd/bases/dscinitialization.opendatahub.io_dscinitializations.yaml b/config/crd/bases/dscinitialization.opendatahub.io_dscinitializations.yaml index 48621d88cc7..a821a61ff36 100644 --- a/config/crd/bases/dscinitialization.opendatahub.io_dscinitializations.yaml +++ b/config/crd/bases/dscinitialization.opendatahub.io_dscinitializations.yaml @@ -132,6 +132,7 @@ spec: field. properties: customCABundle: + default: "" description: A custom CA bundle that will be available for all components in the Data Science Cluster(DSC). This bundle will be stored in odh-trusted-ca-bundle ConfigMap .data.odh-ca-bundle.crt . @@ -147,6 +148,7 @@ spec: pattern: ^(Managed|Unmanaged|Force|Removed)$ type: string required: + - customCABundle - managementState type: object required: diff --git a/config/default/kustomization.yaml b/config/default/kustomization.yaml index fc1b6912725..1a7fe00e8c9 100644 --- a/config/default/kustomization.yaml +++ b/config/default/kustomization.yaml @@ -21,7 +21,7 @@ resources: - ../manager # [WEBHOOK] To enable webhook, uncomment all the sections with [WEBHOOK] prefix including the one in # crd/kustomization.yaml -#- ../webhook +- ../webhook # [CERTMANAGER] To enable cert-manager, uncomment all sections with 'CERTMANAGER'. 'WEBHOOK' components are required. #- ../certmanager # [PROMETHEUS] To enable prometheus monitor, uncomment all sections with 'PROMETHEUS'. @@ -37,6 +37,7 @@ resources: # [WEBHOOK] To enable webhook, uncomment all the sections with [WEBHOOK] prefix including the one in # crd/kustomization.yaml +# Moved below to patches #- manager_webhook_patch.yaml # [CERTMANAGER] To enable cert-manager, uncomment all sections with 'CERTMANAGER'. @@ -76,3 +77,5 @@ resources: patches: - path: manager_auth_proxy_patch.yaml +# [WEBHOOK] +- path: manager_webhook_patch.yaml diff --git a/config/default/manager_webhook_patch.yaml b/config/default/manager_webhook_patch.yaml new file mode 100644 index 00000000000..d76a0133d0d --- /dev/null +++ b/config/default/manager_webhook_patch.yaml @@ -0,0 +1,23 @@ +apiVersion: apps/v1 +kind: Deployment +metadata: + name: controller-manager + namespace: system +spec: + template: + spec: + containers: + - name: manager + ports: + - containerPort: 9443 + name: webhook-server + protocol: TCP + volumeMounts: + - mountPath: /tmp/k8s-webhook-server/serving-certs + name: cert + readOnly: true + volumes: + - name: cert + secret: + defaultMode: 420 + secretName: redhat-ods-operator-controller-webhook-cert diff --git a/config/manifests/bases/rhods-operator.clusterserviceversion.yaml b/config/manifests/bases/rhods-operator.clusterserviceversion.yaml index 6da19c778bf..18f7ec09815 100644 --- a/config/manifests/bases/rhods-operator.clusterserviceversion.yaml +++ b/config/manifests/bases/rhods-operator.clusterserviceversion.yaml @@ -97,6 +97,12 @@ spec: e.g. it provides unified authentication giving a Single Sign On experience. displayName: Service Mesh path: serviceMesh + - description: When set to `Managed`, adds odh-trusted-ca-bundle Configmap to + all namespaces that includes cluster-wide Trusted CA Bundle in .data["ca-bundle.crt"]. + Additionally, this fields allows admins to add custom CA bundles to the + configmap using the .CustomCABundle field. + displayName: Trusted CABundle + path: trustedCABundle - description: Internal development useful field to test customizations. This is not recommended to be used in production environment. displayName: Dev Flags diff --git a/config/webhook/kustomization.yaml b/config/webhook/kustomization.yaml new file mode 100644 index 00000000000..8428859f524 --- /dev/null +++ b/config/webhook/kustomization.yaml @@ -0,0 +1,9 @@ +resources: +- manifests.yaml +- service.yaml + +commonAnnotations: + service.beta.openshift.io/inject-cabundle: "true" + +configurations: +- kustomizeconfig.yaml diff --git a/config/webhook/kustomizeconfig.yaml b/config/webhook/kustomizeconfig.yaml new file mode 100644 index 00000000000..25e21e3c963 --- /dev/null +++ b/config/webhook/kustomizeconfig.yaml @@ -0,0 +1,25 @@ +# the following config is for teaching kustomize where to look at when substituting vars. +# It requires kustomize v2.1.0 or newer to work properly. +nameReference: +- kind: Service + version: v1 + fieldSpecs: + - kind: MutatingWebhookConfiguration + group: admissionregistration.k8s.io + path: webhooks/clientConfig/service/name + - kind: ValidatingWebhookConfiguration + group: admissionregistration.k8s.io + path: webhooks/clientConfig/service/name + +namespace: +- kind: MutatingWebhookConfiguration + group: admissionregistration.k8s.io + path: webhooks/clientConfig/service/namespace + create: true +- kind: ValidatingWebhookConfiguration + group: admissionregistration.k8s.io + path: webhooks/clientConfig/service/namespace + create: true + +varReference: +- path: metadata/annotations diff --git a/config/webhook/manifests.yaml b/config/webhook/manifests.yaml new file mode 100644 index 00000000000..5595314bffb --- /dev/null +++ b/config/webhook/manifests.yaml @@ -0,0 +1,29 @@ +--- +apiVersion: admissionregistration.k8s.io/v1 +kind: ValidatingWebhookConfiguration +metadata: + creationTimestamp: null + name: validating-webhook-configuration +webhooks: +- admissionReviewVersions: + - v1 + clientConfig: + service: + name: webhook-service + namespace: system + path: /validate-opendatahub-io-v1 + failurePolicy: Fail + name: operator.opendatahub.io + rules: + - apiGroups: + - datasciencecluster.opendatahub.io + - dscinitialization.opendatahub.io + apiVersions: + - v1 + operations: + - CREATE + - UPDATE + resources: + - datascienceclusters + - dscinitializations + sideEffects: None diff --git a/config/webhook/service.yaml b/config/webhook/service.yaml new file mode 100644 index 00000000000..72c652ddaae --- /dev/null +++ b/config/webhook/service.yaml @@ -0,0 +1,22 @@ + +apiVersion: v1 +kind: Service +metadata: + labels: + app.kubernetes.io/name: service + app.kubernetes.io/instance: webhook-service + app.kubernetes.io/component: webhook + app.kubernetes.io/created-by: rhods-operator + app.kubernetes.io/part-of: rhods-operator + app.kubernetes.io/managed-by: kustomize + name: webhook-service + namespace: system + annotations: + service.beta.openshift.io/serving-cert-secret-name: redhat-ods-operator-controller-webhook-cert +spec: + ports: + - port: 443 + protocol: TCP + targetPort: 9443 + selector: + control-plane: controller-manager diff --git a/controllers/certconfigmapgenerator/certconfigmapgenerator_controller.go b/controllers/certconfigmapgenerator/certconfigmapgenerator_controller.go index 17732592d39..2b9c94dc182 100644 --- a/controllers/certconfigmapgenerator/certconfigmapgenerator_controller.go +++ b/controllers/certconfigmapgenerator/certconfigmapgenerator_controller.go @@ -49,7 +49,7 @@ func (r *CertConfigmapGeneratorReconciler) SetupWithManager(mgr ctrl.Manager) er // ca bundle in every new namespace created. func (r *CertConfigmapGeneratorReconciler) Reconcile(ctx context.Context, req ctrl.Request) (ctrl.Result, error) { // Request includes namespace that is newly created or where odh-trusted-ca-bundle configmap is updated. - r.Log.Info("Reconciling certConfigMapGenerator.", " CertConfigMapGenerator Request.Namespace", req.NamespacedName) + r.Log.Info("Reconciling certConfigMapGenerator.", "CertConfigMapGenerator Request.Namespace", req.NamespacedName) // Get namespace instance userNamespace := &corev1.Namespace{} err := r.Client.Get(ctx, client.ObjectKey{Name: req.Namespace}, userNamespace) @@ -71,9 +71,6 @@ func (r *CertConfigmapGeneratorReconciler) Reconcile(ctx context.Context, req ct return ctrl.Result{}, nil case 1: dsciInstance = &dsciInstances.Items[0] - default: - message := "only one instance of DSCInitialization object is allowed" - return ctrl.Result{}, errors.New(message) } if dsciInstance.Spec.TrustedCABundle.ManagementState != operatorv1.Managed { diff --git a/controllers/datasciencecluster/datasciencecluster_controller.go b/controllers/datasciencecluster/datasciencecluster_controller.go index 3010faa80f8..72e119636f4 100644 --- a/controllers/datasciencecluster/datasciencecluster_controller.go +++ b/controllers/datasciencecluster/datasciencecluster_controller.go @@ -82,7 +82,7 @@ const ( // Reconcile is part of the main kubernetes reconciliation loop which aims to // move the current state of the cluster closer to the desired state. -func (r *DataScienceClusterReconciler) Reconcile(ctx context.Context, req ctrl.Request) (ctrl.Result, error) { //nolint:gocyclo,maintidx +func (r *DataScienceClusterReconciler) Reconcile(ctx context.Context, req ctrl.Request) (ctrl.Result, error) { //nolint:gocyclo r.Log.Info("Reconciling DataScienceCluster resources", "Request.Name", req.Name) instances := &dsc.DataScienceClusterList{} @@ -138,19 +138,6 @@ func (r *DataScienceClusterReconciler) Reconcile(ctx context.Context, req ctrl.R return reconcile.Result{Requeue: true}, nil } - if len(instances.Items) > 1 { - message := fmt.Sprintf("only one instance of DataScienceCluster object is allowed. Update existing instance %s", req.Name) - err := errors.New(message) - _ = r.reportError(err, instance, message) - - _, _ = r.updateStatus(ctx, instance, func(saved *dsc.DataScienceCluster) { - status.SetErrorCondition(&saved.Status.Conditions, status.DuplicateDataScienceCluster, message) - saved.Status.Phase = status.PhaseError - }) - - return ctrl.Result{}, err - } - // Verify a valid DSCInitialization instance is created dsciInstances := &dsci.DSCInitializationList{} err = r.Client.List(ctx, dsciInstances) @@ -161,7 +148,7 @@ func (r *DataScienceClusterReconciler) Reconcile(ctx context.Context, req ctrl.R } // Update phase to error state if DataScienceCluster is created without valid DSCInitialization - switch len(dsciInstances.Items) { + switch len(dsciInstances.Items) { // only handle number as 0 or 1, others won't be existed since webhook block creation case 0: reason := status.ReconcileFailed message := "Failed to get a valid DSCInitialization instance" @@ -177,13 +164,6 @@ func (r *DataScienceClusterReconciler) Reconcile(ctx context.Context, req ctrl.R case 1: dscInitializationSpec := dsciInstances.Items[0].Spec dscInitializationSpec.DeepCopyInto(r.DataScienceCluster.DSCISpec) - default: - message := "only one instance of DSCInitialization object is allowed" - _, _ = r.updateStatus(ctx, instance, func(saved *dsc.DataScienceCluster) { - status.SetErrorCondition(&saved.Status.Conditions, status.DuplicateDSCInitialization, message) - saved.Status.Phase = status.PhaseError - }) - return ctrl.Result{}, errors.New(message) } if instance.ObjectMeta.DeletionTimestamp.IsZero() { diff --git a/controllers/dscinitialization/dscinitialization_controller.go b/controllers/dscinitialization/dscinitialization_controller.go index be34f146816..cd076d5cee2 100644 --- a/controllers/dscinitialization/dscinitialization_controller.go +++ b/controllers/dscinitialization/dscinitialization_controller.go @@ -19,8 +19,6 @@ package dscinitialization import ( "context" - "errors" - "fmt" "path/filepath" "reflect" @@ -89,33 +87,11 @@ func (r *DSCInitializationReconciler) Reconcile(ctx context.Context, req ctrl.Re } var instance *dsciv1.DSCInitialization - switch { + switch { // only handle number as 0 or 1, others won't be existed since webhook block creation case len(instances.Items) == 0: return ctrl.Result{}, nil case len(instances.Items) == 1: instance = &instances.Items[0] - case len(instances.Items) > 1: - // find out the one by created timestamp and use it as the default one - earliestDSCI := &instances.Items[0] - for _, instance := range instances.Items { - currentDSCI := instance - if currentDSCI.CreationTimestamp.Before(&earliestDSCI.CreationTimestamp) { - earliestDSCI = ¤tDSCI - } - } - message := fmt.Sprintf("only one instance of DSCInitialization object is allowed. Please delete other instances than %s", earliestDSCI.Name) - // update all instances Message and Status - for _, deletionInstance := range instances.Items { - deletionInstance := deletionInstance - if deletionInstance.Name != earliestDSCI.Name { - _, _ = r.updateStatus(ctx, &deletionInstance, func(saved *dsciv1.DSCInitialization) { - status.SetErrorCondition(&saved.Status.Conditions, status.DuplicateDSCInitialization, message) - saved.Status.Phase = status.PhaseError - }) - } - } - - return ctrl.Result{}, errors.New(message) } if instance.ObjectMeta.DeletionTimestamp.IsZero() { diff --git a/controllers/dscinitialization/dscinitialization_test.go b/controllers/dscinitialization/dscinitialization_test.go index 6a726a1998b..17b510c4d2c 100644 --- a/controllers/dscinitialization/dscinitialization_test.go +++ b/controllers/dscinitialization/dscinitialization_test.go @@ -20,6 +20,7 @@ import ( const ( workingNamespace = "test-operator-ns" + applicationName = "default-dsci" applicationNamespace = "test-application-ns" configmapName = "odh-common-config" monitoringNamespace = "test-monitoring-ns" @@ -29,10 +30,10 @@ const ( var _ = Describe("DataScienceCluster initialization", func() { Context("Creation of related resources", func() { // must be default as instance name, or it will break - const applicationName = "default-dsci" + BeforeEach(func() { // when - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) + desiredDsci := createDSCI(operatorv1.Managed, operatorv1.Managed, monitoringNamespace) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -119,7 +120,7 @@ var _ = Describe("DataScienceCluster initialization", func() { const applicationName = "default-dsci" It("Should not create monitoring namespace if monitoring is disabled", func() { // when - desiredDsci := createDSCI(applicationName, operatorv1.Removed, operatorv1.Managed, monitoringNamespace2) + desiredDsci := createDSCI(operatorv1.Removed, operatorv1.Managed, monitoringNamespace2) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -135,7 +136,7 @@ var _ = Describe("DataScienceCluster initialization", func() { }) It("Should create default monitoring namespace if monitoring enabled", func() { // when - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace2) + desiredDsci := createDSCI(operatorv1.Managed, operatorv1.Managed, monitoringNamespace2) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -154,22 +155,6 @@ var _ = Describe("DataScienceCluster initialization", func() { Context("Handling existing resources", func() { AfterEach(cleanupResources) - const applicationName = "default-dsci" - - It("Should not have more than one DSCI instance in the cluster", func() { - - anotherApplicationName := "default2" - // given - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) - Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) - // when - desiredDsci2 := createDSCI(anotherApplicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) - // then - Eventually(dscInitializationIsReady(anotherApplicationName, workingNamespace, desiredDsci2)). - WithTimeout(timeout). - WithPolling(interval). - Should(BeFalse()) - }) It("Should not update rolebinding if it exists", func() { applicationName := envtestutil.AppendRandomNameTo("rolebinding-test") @@ -199,7 +184,7 @@ var _ = Describe("DataScienceCluster initialization", func() { Should(BeTrue()) // when - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) + desiredDsci := createDSCI(operatorv1.Managed, operatorv1.Managed, monitoringNamespace) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -240,7 +225,7 @@ var _ = Describe("DataScienceCluster initialization", func() { Should(BeTrue()) // when - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) + desiredDsci := createDSCI(operatorv1.Managed, operatorv1.Managed, monitoringNamespace) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -277,7 +262,7 @@ var _ = Describe("DataScienceCluster initialization", func() { Should(BeTrue()) // when - desiredDsci := createDSCI(applicationName, operatorv1.Managed, operatorv1.Managed, monitoringNamespace) + desiredDsci := createDSCI(operatorv1.Managed, operatorv1.Managed, monitoringNamespace) Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) foundDsci := &dsci.DSCInitialization{} Eventually(dscInitializationIsReady(applicationName, workingNamespace, foundDsci)). @@ -349,14 +334,14 @@ func objectExists(ns string, name string, obj client.Object) func() bool { //nol } } -func createDSCI(appName string, enableMonitoring operatorv1.ManagementState, enableTrustedCABundle operatorv1.ManagementState, monitoringNS string) *dsci.DSCInitialization { +func createDSCI(enableMonitoring operatorv1.ManagementState, enableTrustedCABundle operatorv1.ManagementState, monitoringNS string) *dsci.DSCInitialization { return &dsci.DSCInitialization{ TypeMeta: metav1.TypeMeta{ Kind: "DSCInitialization", APIVersion: "v1", }, ObjectMeta: metav1.ObjectMeta{ - Name: appName, + Name: applicationName, Namespace: workingNamespace, }, Spec: dsci.DSCInitializationSpec{ diff --git a/controllers/webhook/webhook.go b/controllers/webhook/webhook.go new file mode 100644 index 00000000000..f730dc78703 --- /dev/null +++ b/controllers/webhook/webhook.go @@ -0,0 +1,111 @@ +/* +Copyright 2023. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package webhook + +import ( + "context" + "fmt" + "net/http" + + admissionv1 "k8s.io/api/admission/v1" + "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" + "k8s.io/apimachinery/pkg/runtime/schema" + ctrl "sigs.k8s.io/controller-runtime" + "sigs.k8s.io/controller-runtime/pkg/client" + "sigs.k8s.io/controller-runtime/pkg/webhook" + "sigs.k8s.io/controller-runtime/pkg/webhook/admission" +) + +var log = ctrl.Log.WithName("odh-controller-webhook") + +//+kubebuilder:webhook:path=/validate-opendatahub-io-v1,mutating=false,failurePolicy=fail,sideEffects=None,groups=datasciencecluster.opendatahub.io;dscinitialization.opendatahub.io,resources=datascienceclusters;dscinitializations,verbs=create;update,versions=v1,name=operator.opendatahub.io,admissionReviewVersions=v1 +//nolint:lll + +type OpenDataHubWebhook struct { + client client.Client + decoder *admission.Decoder +} + +func (w *OpenDataHubWebhook) SetupWithManager(mgr ctrl.Manager) { + hookServer := mgr.GetWebhookServer() + odhWebhook := &webhook.Admission{ + Handler: w, + } + hookServer.Register("/validate-opendatahub-io-v1", odhWebhook) +} + +func (w *OpenDataHubWebhook) InjectDecoder(d *admission.Decoder) error { + w.decoder = d + return nil +} + +func (w *OpenDataHubWebhook) InjectClient(c client.Client) error { + w.client = c + return nil +} + +func (w *OpenDataHubWebhook) checkDupCreation(ctx context.Context, req admission.Request) admission.Response { + if req.Operation != admissionv1.Create { + return admission.Allowed(fmt.Sprintf("duplication check: skipping %v request", req.Operation)) + } + + switch req.Kind.Kind { + case "DataScienceCluster": + case "DSCInitialization": + default: + log.Info("Got wrong kind", "kind", req.Kind.Kind) + return admission.Errored(http.StatusBadRequest, nil) + } + + gvk := schema.GroupVersionKind{ + Group: req.Kind.Group, + Version: req.Kind.Version, + Kind: req.Kind.Kind, + } + + list := &unstructured.UnstructuredList{} + list.SetGroupVersionKind(gvk) + + if err := w.client.List(ctx, list); err != nil { + return admission.Errored(http.StatusBadRequest, err) + } + + // if len == 1 now creation of #2 is being handled + if len(list.Items) > 0 { + return admission.Denied(fmt.Sprintf("Only one instance of %s object is allowed", req.Kind.Kind)) + } + + return admission.Allowed(fmt.Sprintf("%s duplication check passed", req.Kind.Kind)) +} + +func (w *OpenDataHubWebhook) Handle(ctx context.Context, req admission.Request) admission.Response { + var resp admission.Response + + // Handle only Create and Update + if req.Operation == admissionv1.Delete || req.Operation == admissionv1.Connect { + msg := fmt.Sprintf("ODH skipping %v request", req.Operation) + log.Info(msg) + return admission.Allowed(msg) + } + + resp = w.checkDupCreation(ctx, req) + if !resp.Allowed { + return resp + } + + return admission.Allowed(fmt.Sprintf("%s allowed", req.Kind.Kind)) +} diff --git a/controllers/webhook/webhook_suite_test.go b/controllers/webhook/webhook_suite_test.go new file mode 100644 index 00000000000..95ce437c10c --- /dev/null +++ b/controllers/webhook/webhook_suite_test.go @@ -0,0 +1,244 @@ +/* +Copyright 2023. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package webhook_test + +import ( + "context" + "crypto/tls" + "fmt" + "net" + "path/filepath" + "testing" + "time" + + operatorv1 "github.com/openshift/api/operator/v1" + admissionv1beta1 "k8s.io/api/admission/v1beta1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" + "k8s.io/apimachinery/pkg/runtime" + "k8s.io/client-go/rest" + ctrl "sigs.k8s.io/controller-runtime" + "sigs.k8s.io/controller-runtime/pkg/client" + "sigs.k8s.io/controller-runtime/pkg/envtest" + logf "sigs.k8s.io/controller-runtime/pkg/log" + "sigs.k8s.io/controller-runtime/pkg/log/zap" + + dsc "github.com/opendatahub-io/opendatahub-operator/v2/apis/datasciencecluster/v1" + dsci "github.com/opendatahub-io/opendatahub-operator/v2/apis/dscinitialization/v1" + "github.com/opendatahub-io/opendatahub-operator/v2/components" + "github.com/opendatahub-io/opendatahub-operator/v2/components/codeflare" + "github.com/opendatahub-io/opendatahub-operator/v2/components/dashboard" + "github.com/opendatahub-io/opendatahub-operator/v2/components/datasciencepipelines" + "github.com/opendatahub-io/opendatahub-operator/v2/components/kserve" + "github.com/opendatahub-io/opendatahub-operator/v2/components/modelmeshserving" + "github.com/opendatahub-io/opendatahub-operator/v2/components/ray" + "github.com/opendatahub-io/opendatahub-operator/v2/components/trustyai" + "github.com/opendatahub-io/opendatahub-operator/v2/components/workbenches" + "github.com/opendatahub-io/opendatahub-operator/v2/controllers/webhook" + + . "github.com/onsi/ginkgo/v2" + . "github.com/onsi/gomega" +) + +const ( + namespace = "webhook-test-ns" + nameBase = "webhook-test" +) + +// These tests use Ginkgo (BDD-style Go testing framework). Refer to +// http://onsi.github.io/ginkgo/ to learn more about Ginkgo. + +var cfg *rest.Config +var k8sClient client.Client +var testEnv *envtest.Environment +var ctx context.Context +var cancel context.CancelFunc + +func TestAPIs(t *testing.T) { + RegisterFailHandler(Fail) + + RunSpecs(t, "Webhook Suite") +} + +var _ = BeforeSuite(func() { + logf.SetLogger(zap.New(zap.WriteTo(GinkgoWriter), zap.UseDevMode(true))) + + ctx, cancel = context.WithCancel(context.TODO()) + + By("bootstrapping test environment") + testEnv = &envtest.Environment{ + CRDDirectoryPaths: []string{filepath.Join("..", "..", "config", "crd", "bases")}, + ErrorIfCRDPathMissing: false, + WebhookInstallOptions: envtest.WebhookInstallOptions{ + Paths: []string{filepath.Join("..", "..", "config", "webhook")}, + }, + } + + var err error + // cfg is defined in this file globally. + cfg, err = testEnv.Start() + Expect(err).NotTo(HaveOccurred()) + Expect(cfg).NotTo(BeNil()) + + scheme := runtime.NewScheme() + // DSCI + err = dsci.AddToScheme(scheme) + Expect(err).NotTo(HaveOccurred()) + // DSC + err = dsc.AddToScheme(scheme) + Expect(err).NotTo(HaveOccurred()) + // Webhook + err = admissionv1beta1.AddToScheme(scheme) + Expect(err).NotTo(HaveOccurred()) + + k8sClient, err = client.New(cfg, client.Options{Scheme: scheme}) + Expect(err).NotTo(HaveOccurred()) + Expect(k8sClient).NotTo(BeNil()) + + // start webhook server using Manager + webhookInstallOptions := &testEnv.WebhookInstallOptions + mgr, err := ctrl.NewManager(cfg, ctrl.Options{ + Scheme: scheme, + Host: webhookInstallOptions.LocalServingHost, + Port: webhookInstallOptions.LocalServingPort, + CertDir: webhookInstallOptions.LocalServingCertDir, + LeaderElection: false, + MetricsBindAddress: "0", + }) + Expect(err).NotTo(HaveOccurred()) + + (&webhook.OpenDataHubWebhook{}).SetupWithManager(mgr) + + //+kubebuilder:scaffold:webhook + + go func() { + defer GinkgoRecover() + err = mgr.Start(ctx) + Expect(err).NotTo(HaveOccurred()) + }() + + // wait for the webhook server to get ready + dialer := &net.Dialer{Timeout: time.Second} + addrPort := fmt.Sprintf("%s:%d", webhookInstallOptions.LocalServingHost, webhookInstallOptions.LocalServingPort) + Eventually(func() error { + conn, err := tls.DialWithDialer(dialer, "tcp", addrPort, &tls.Config{InsecureSkipVerify: true}) //nolint:gosec + if err != nil { + return err + } + conn.Close() + return nil + }).Should(Succeed()) + +}) + +var _ = AfterSuite(func() { + cancel() + By("tearing down the test environment") + err := testEnv.Stop() + Expect(err).NotTo(HaveOccurred()) +}) + +var _ = Describe("DSC/DSCI webhook", func() { + It("Should not have more than one DSCI instance in the cluster", func() { + desiredDsci := newDSCI(nameBase + "-dsci-1") + Expect(k8sClient.Create(context.Background(), desiredDsci)).Should(Succeed()) + desiredDsci2 := newDSCI(nameBase + "-dsci-2") + Expect(k8sClient.Create(context.Background(), desiredDsci2)).ShouldNot(Succeed()) + }) + + It("Should block creation of second DSC instance", func() { + dscSpec := newDSC(nameBase+"-dsc-1", namespace) + Expect(k8sClient.Create(context.Background(), dscSpec)).Should(Succeed()) + dscSpec = newDSC(nameBase+"-dsc-2", namespace) + Expect(k8sClient.Create(context.Background(), dscSpec)).ShouldNot(Succeed()) + }) +}) + +func newDSCI(appName string) *dsci.DSCInitialization { + monitoringNS := "monitoring-namespace" + return &dsci.DSCInitialization{ + TypeMeta: metav1.TypeMeta{ + Kind: "DSCInitialization", + APIVersion: "v1", + }, + ObjectMeta: metav1.ObjectMeta{ + Name: appName, + Namespace: namespace, + }, + Spec: dsci.DSCInitializationSpec{ + ApplicationsNamespace: namespace, + Monitoring: dsci.Monitoring{ + Namespace: monitoringNS, + ManagementState: operatorv1.Managed, + }, + TrustedCABundle: dsci.TrustedCABundleSpec{ + ManagementState: operatorv1.Managed, + }, + }, + } +} +func newDSC(name string, namespace string) *dsc.DataScienceCluster { + return &dsc.DataScienceCluster{ + ObjectMeta: metav1.ObjectMeta{ + Name: name, + Namespace: namespace, + }, + Spec: dsc.DataScienceClusterSpec{ + Components: dsc.Components{ + Dashboard: dashboard.Dashboard{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + Workbenches: workbenches.Workbenches{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + ModelMeshServing: modelmeshserving.ModelMeshServing{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + DataSciencePipelines: datasciencepipelines.DataSciencePipelines{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + Kserve: kserve.Kserve{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + CodeFlare: codeflare.CodeFlare{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + Ray: ray.Ray{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + TrustyAI: trustyai.TrustyAI{ + Component: components.Component{ + ManagementState: operatorv1.Removed, + }, + }, + }, + }, + } +} diff --git a/main.go b/main.go index 67f46d3b772..96ba5af5a94 100644 --- a/main.go +++ b/main.go @@ -17,6 +17,7 @@ limitations under the License. package main import ( + "context" "flag" "os" @@ -45,6 +46,7 @@ import ( "sigs.k8s.io/controller-runtime/pkg/client/config" "sigs.k8s.io/controller-runtime/pkg/healthz" "sigs.k8s.io/controller-runtime/pkg/log/zap" + "sigs.k8s.io/controller-runtime/pkg/manager" kfdefv1 "github.com/opendatahub-io/opendatahub-operator/apis/kfdef.apps.kubeflow.org/v1" dsc "github.com/opendatahub-io/opendatahub-operator/v2/apis/datasciencecluster/v1" @@ -54,6 +56,7 @@ import ( datascienceclustercontrollers "github.com/opendatahub-io/opendatahub-operator/v2/controllers/datasciencecluster" dscicontr "github.com/opendatahub-io/opendatahub-operator/v2/controllers/dscinitialization" "github.com/opendatahub-io/opendatahub-operator/v2/controllers/secretgenerator" + "github.com/opendatahub-io/opendatahub-operator/v2/controllers/webhook" "github.com/opendatahub-io/opendatahub-operator/v2/pkg/deploy" "github.com/opendatahub-io/opendatahub-operator/v2/pkg/upgrade" ) @@ -137,6 +140,8 @@ func main() { os.Exit(1) } + (&webhook.OpenDataHubWebhook{}).SetupWithManager(mgr) + if err = (&dscicontr.DSCInitializationReconciler{ Client: mgr.GetClient(), Scheme: mgr.GetScheme(), @@ -203,8 +208,18 @@ func main() { // Check if user opted for disabling DSC configuration _, disableDSCConfig := os.LookupEnv("DISABLE_DSC_CONFIG") if !disableDSCConfig { - if err = upgrade.CreateDefaultDSCI(setupClient, platform, dscApplicationsNamespace, dscMonitoringNamespace); err != nil { - setupLog.Error(err, "unable to create initial setup for the operator") + var createDefaultDSCIFunc manager.RunnableFunc = func(ctx context.Context) error { + err := upgrade.CreateDefaultDSCI(setupClient, platform, dscApplicationsNamespace, dscMonitoringNamespace) + if err != nil { + setupLog.Error(err, "unable to create initial setup for the operator") + } + return err + } + + err := mgr.Add(createDefaultDSCIFunc) + if err != nil { + setupLog.Error(err, "error scheduling DSCI creation") + os.Exit(1) } } diff --git a/tests/e2e/controller_setup_test.go b/tests/e2e/controller_setup_test.go index bb005142fce..b5f058c4957 100644 --- a/tests/e2e/controller_setup_test.go +++ b/tests/e2e/controller_setup_test.go @@ -76,9 +76,9 @@ func NewTestContext() (*testContext, error) { //nolint:golint,revive // Only use } // setup DSCI CR since we do not create automatically by operator - testDSCI := setupDSCICR() + testDSCI := setupDSCICR("e2e-test-dsci") // Setup DataScienceCluster CR - testDSC := setupDSCInstance() + testDSC := setupDSCInstance("e2e-test") return &testContext{ cfg: config, diff --git a/tests/e2e/dsc_creation_test.go b/tests/e2e/dsc_creation_test.go index b5fcc71b888..5c2e078f78c 100644 --- a/tests/e2e/dsc_creation_test.go +++ b/tests/e2e/dsc_creation_test.go @@ -13,6 +13,9 @@ import ( autoscalingv1 "k8s.io/api/autoscaling/v1" "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" + "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" + "k8s.io/apimachinery/pkg/runtime" + "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/types" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/client-go/util/retry" @@ -34,10 +37,16 @@ func creationTestSuite(t *testing.T) { err = testCtx.testDSCICreation() require.NoError(t, err, "error creating DSCI CR") }) + t.Run("Creation of more than one of DSCInitialization instance", func(t *testing.T) { + testCtx.testDSCIDuplication(t) + }) t.Run("Creation of DataScienceCluster instance", func(t *testing.T) { err = testCtx.testDSCCreation() require.NoError(t, err, "error creating DataScienceCluster instance") }) + t.Run("Creation of more than one of DataScienceCluster instance", func(t *testing.T) { + testCtx.testDSCDuplication(t) + }) t.Run("Validate all deployed components", func(t *testing.T) { err = testCtx.testAllApplicationCreation(t) require.NoError(t, err, "error testing deployments for DataScienceCluster: "+testCtx.testDsc.Name) @@ -132,6 +141,56 @@ func (tc *testContext) testDSCCreation() error { return nil } +func (tc *testContext) requireInstalled(t *testing.T, gvk schema.GroupVersionKind) { + t.Helper() + list := &unstructured.UnstructuredList{} + list.SetGroupVersionKind(gvk) + + err := tc.customClient.List(tc.ctx, list) + require.NoErrorf(t, err, "Could not get %s list", gvk.Kind) + + require.Greaterf(t, len(list.Items), 0, "%s has not been installed", gvk.Kind) +} + +func (tc *testContext) testDuplication(t *testing.T, gvk schema.GroupVersionKind, o any) { + t.Helper() + tc.requireInstalled(t, gvk) + + u, err := runtime.DefaultUnstructuredConverter.ToUnstructured(o) + require.NoErrorf(t, err, "Could not unstructure %s", gvk.Kind) + + obj := &unstructured.Unstructured{ + Object: u, + } + obj.SetGroupVersionKind(gvk) + + err = tc.customClient.Create(tc.ctx, obj) + + require.Errorf(t, err, "Could create second %s", gvk.Kind) +} + +func (tc *testContext) testDSCIDuplication(t *testing.T) { //nolint:thelper + gvk := schema.GroupVersionKind{ + Group: "dscinitialization.opendatahub.io", + Version: "v1", + Kind: "DSCInitialization", + } + dup := setupDSCICR("e2e-test-dsci-dup") + + tc.testDuplication(t, gvk, dup) +} + +func (tc *testContext) testDSCDuplication(t *testing.T) { //nolint:thelper + gvk := schema.GroupVersionKind{ + Group: "datasciencecluster.opendatahub.io", + Version: "v1", + Kind: "DataScienceCluster", + } + dup := setupDSCInstance("e2e-test-dup") + + tc.testDuplication(t, gvk, dup) +} + func (tc *testContext) testAllApplicationCreation(t *testing.T) error { //nolint:funlen,thelper // Validate test instance is in Ready state diff --git a/tests/e2e/helper_test.go b/tests/e2e/helper_test.go index 592a95cff4b..c426d13feeb 100644 --- a/tests/e2e/helper_test.go +++ b/tests/e2e/helper_test.go @@ -53,10 +53,10 @@ func (tc *testContext) waitForControllerDeployment(name string, replicas int32) return err } -func setupDSCICR() *dsci.DSCInitialization { +func setupDSCICR(name string) *dsci.DSCInitialization { dsciTest := &dsci.DSCInitialization{ ObjectMeta: metav1.ObjectMeta{ - Name: "e2e-test-dsci", + Name: name, }, Spec: dsci.DSCInitializationSpec{ ApplicationsNamespace: "opendatahub", @@ -73,10 +73,10 @@ func setupDSCICR() *dsci.DSCInitialization { return dsciTest } -func setupDSCInstance() *dsc.DataScienceCluster { +func setupDSCInstance(name string) *dsc.DataScienceCluster { dscTest := &dsc.DataScienceCluster{ ObjectMeta: metav1.ObjectMeta{ - Name: "e2e-test-dsc", + Name: name, }, Spec: dsc.DataScienceClusterSpec{ Components: dsc.Components{