deployment

No menu items for this category

On Premises Kubernetes Deployment

OpenMetadata supports the Installation and Running of application on OnPremises Kubernetes through Helm Charts. However, there are some additional configurations which needs to be done as prerequisites for the same.

This guide presumes you have an on premises Kubernetes cluster setup, and you are installing OpenMetadata in default namespace.

We support

  • MySQL engine version 8 or higher
  • PostgreSQL engine version 12 or higher
  • ElasticSearch version 8.X (upto 8.10.2) or OpenSearch Version 2.X (upto 2.7)

Once you have the External Database and Search Engine configured, you can update the environment variables below for OpenMetadata kubernetes deployments to connect with Database and ElasticSearch.

Make sure to create database and search engine credentials as Kubernetes Secrets mentioned here.

Also, disable MySQL and ElasticSearch from OpenMetadata Dependencies Helm Charts as mentioned in the FAQs here.

OpenMetadata helm chart depends on Airflow and Airflow expects a persistent disk that support ReadWriteMany (the volume can be mounted as read-write by many nodes).

The workaround is to create nfs-share and use that as the persistent claim to deploy OpenMetadata by implementing the following steps in order.

This guide assumes you have NFS Server already setup with Hostname or IP Address which is reachable from your on premises Kubernetes cluster, and you have configured a path to be used for OpenMetadata Airflow Helm Dependency.

To provision PersistentVolume dynamically using the StorageClass, you need to install the NFS provisioner. It is recommended to use nfs-subdir-external-provisioner helm charts for this case.

Replace the NFS_HOSTNAME_OR_IP with your NFS Server value and run the commands.

This will create a new StorageClass with nfs-subdir-external-provisioner. You can view the same using the kubectl command kubectl get storageclass -n nfs-provisioner.

Create Persistent Volumes and Persistent Volume claims with the below command.

Create Persistent Volumes and Persistent Volume claims with the below command.

Since airflow pods run as non-root users, they would not have write access on the nfs server volumes. In order to fix the permission here, spin up a pod with persistent volumes attached and run it once.

Airflow runs the pods with linux username as airflow and linux user id as 50000.

Run the below command to create the pod and fix the permissions

Override openmetadata dependencies airflow helm values to bind the nfs persistent volumes for DAGs and logs.

For more information on airflow helm chart values, please refer to airflow-helm.

When deploying openmetadata dependencies helm chart, use the below command -

The above command uses configurations defined here. You can modify any configuration and deploy by passing your own values.yaml

Once the openmetadata dependencies helm chart deployed, you can then run the below command to install the openmetadata helm chart -

Again, this uses the values defined here. Use the --values flag to point to your own YAML configuration if needed.