No menu items for this category

Upgrade on Kubernetes

This guide will help you upgrade your OpenMetadata Kubernetes Application with automated helm hooks.

This guide assumes that you have an OpenMetadata deployment that you installed and configured following the Kubernetes Deployment guide.

We also assume that your helm chart release names are openmetadata and openmetadata-dependencies and namespace used is default.


Everytime that you plan on upgrading OpenMetadata to a newer version, make sure to go over all these steps:

Before upgrading your OpenMetadata version we strongly recommend backing up the metadata.

The source of truth is stored in the underlying database (MySQL and Postgres supported). During each version upgrade there is a database migration process that needs to run. It will directly attack your database and update the shape of the data to the newest OpenMetadata release.

It is important that we backup the data because if we face any unexpected issues during the upgrade process, you will be able to get back to the previous version without any loss.

You can learn more about how the migration process works here.

  • To run the backup and restore commands, please make sure that you are always in the latest openmetadata-ingestion version to have all the improvements shipped in the CLI.
  • Also, make sure you have connectivity between your database (MySQL / PostgreSQL) and the host machine where you will be running the below commands.

1. Create a Virtual Environment and Install the Backup CLI

Validate the installed metadata version with python -m metadata --version

2. Run the Backup

If using MySQL:

If using Postgres:

3. Store the backup file somewhere safe

The above command will generate a backup file with extension as .sql. You can copy the name from the backup command output.

Make sure to store it somewhere safe in case you need to restore the data later.

You can refer to the following guide to get more details about the backup and restore:

Before running the migrations, it is important to update these parameters to ensure there are no runtime errors. A safe value would be setting them to 20MB.

If using MySQL

You can update it via SQL (note that it will reset after the server restarts):

To make the configuration persistent, you'd need to navigate to your MySQL Server install directory and update the my.ini or my.cnf files with sort_buffer_size = 20971520.

If using RDS, you will need to update your instance's Parameter Group to include the above change.

If using Postgres

You can update it via SQL (not that it will reset after the server restarts):

To make the configuration persistent, you'll need to update the postgresql.conf file with work_mem = 20MB.

If using RDS, you will need to update your instance's Parameter Group to include the above change.

Note that this value would depend on the size of your query_entity table. If you still see Out of Sort Memory Errors during the migration after bumping this value, you can increase them further.

After the migration is finished, you can revert this changes.

Deprecation Notice

  • OpenMetadata only supports Python version 3.8 to 3.10. We will add support for 3.11 in the release 1.3.
  • OpenMetadata version 0.13.x is deprecated.

Breaking Changes

The following configuration block has been removed from openmetadata.yaml:

This change removes the traditional way of providing Custom URL logos configurations as part of OpenMetadata Configurations file and migrate this to be driven and configured right from UI from Settings > OpenMetadata > Custom Logo.

The same applies to the Login Configuration, which can now be configured under Settings > OpenMetadata > Login Configuration.

Note that these environment variables will now have no effect. If you are deploying on Bare Metal, make sure to use the latest openmetadata.yaml file.

As part of 1.2.1, we migrated the base dependencies for OpenMetadata Helm Chart to use OpenSearch version 2.7 instead of ElasticSearch 8.X. This is a reactive change done as community driven ElasticSearch Helm Chart project has been deprecated in the favor of Elastic Stack Operator which cannot be added as an helm chart dependency.

For new users, this is an unnoticeable change who will be installing the OpenMetadata dependencies using quickstart guides.

For existing users, who have their proof-of-concept environments using the OpenMetadata Dependencies and are looking to upgrade to newer helm release -

  • The default OpenMetadata helm values for openmetadata.config.elasticsearch.* has been updated to connect to OpenSearch from OpenMetadata Dependencies Helm Chart. Please refer to the helm values and update your custom installation accordingly.
  • Post upgrade, you will need to follow the steps here to rebuild and reindex your search indexing.

With 1.2.X, the environment variable DB_USE_SSL is deprecated in favour of DB_PARAMS. For Bare Metal and Docker Deployment, Add / Update the variable DB_PARAMS to allowPublicKeyRetrieval=true&useSSL=true&serverTimezone=UTC to enable ssl security to connect to database. For Kubernetes Deployment, openmetadata.config.database.dbParams is available to pass the above values as helm values.

  • The OpenMetadata Server is now based on JDK 17
  • OpenMetadata now requires Elasticsearch version 8.10.2 or Opensearch version 2.7

There is no direct migration to bump the indexes to the new supported versions. You might see errors like:

In order to move forward, you must remove the volumes or delete the indexes directly from your search instances. Note that OpenMetadata stores everything in the database, so indexes can be recreated from the UI. We will show you how in the Post-Upgrade Steps.

  • Added a new key openmetadata.config.database.dbParams to pass extra database parameters as string format, e.g., useSSL=true&serverTimezone=UTC.
  • Removed the entry for openmetadata.config.database.dbUseSSL. You should use openmetadata.config.database.dbParams instead.
  • Updated the ElasticSearch Helm Chart Dependencies to version 8.5.1

The Query Entity now has the service property, linking the Query to the Database Service that it belongs to. Note that service is a required property both for the Query Entity and the Create Query Entity.

During the migrations, we pick up the service from the tables from queryUsedIn. If this information is not available, then there is no way to link a query to a service and the query will be removed.

  • Domo Database, Dashboard and Pipeline renamed the sandboxDomain in favor of instanceDomain.
  • The DatabaseMetadata configuration renamed viewParsingTimeoutLimit to queryParsingTimeoutLimit.
  • The DatabaseMetadata configuration removed the markAllDeletedTables option. For simplicity, we'll only mark as deleted the tables coming from the filtered ingestion results.

We have reorganized the structure of the Workflow classes, which requires updated imports:

  • Metadata Workflow

    • From: from metadata.ingestion.api.workflow import Workflow
    • To: from metadata.workflow.metadata import MetadataWorkflow
  • Lineage Workflow

    • From: from metadata.ingestion.api.workflow import Workflow
    • To: from metadata.workflow.metadata import MetadataWorkflow (same as metadata)
  • Usage Workflow

    • From: from metadata.ingestion.api.workflow import Workflow
    • To: from metadata.workflow.usage import UsageWorkflow
  • Profiler Workflow

    • From: from metadata.profiler.api.workflow import ProfilerWorkflow
    • To: from metadata.workflow.profiler import ProfilerWorkflow
  • Data Quality Workflow

    • From: from metadata.data_quality.api.workflow import TestSuiteWorkflow
    • To: from metadata.workflow.data_quality import TestSuiteWorkflow
  • Data Insights Workflow

    • From: from metadata.data_insight.api.workflow import DataInsightWorkflow
    • To: from metadata.workflow.data_insight import DataInsightWorkflow
  • Elasticsearch Reindex Workflow

    • From: from metadata.ingestion.api.workflow import Workflow
    • To: from metadata.workflow.metadata import MetadataWorkflow (same as metadata)

The Workflow class that you import can then be called as follows:

If you try to run your workflows externally and start noticing ImportErrors, you will need to review the points above.

In 1.1.7 and below you could run the Usage Workflow as metadata ingest -c <path to yaml>. Now, the Usage Workflow has its own command metadata usage -c <path to yaml>.

In 1.2.0 we have reorganized the internals of our Workflow handling to centralize status & exception management. This will simplify how you need to take care of status and exceptions on your Custom Connectors code, while helping developers to make decisions on those errors that need to be shared in the Workflow.

If you want to take a look at an updated Custom Connector and its changes, you can review the demo PR.

Let's list the changes down:

  1. You don't need to handle the SourceStatus anymore. The new basic Workflow class will take care of things for you. Therefore, this import from metadata.ingestion.api.source import SourceStatus is deprecated.
  2. The Source class is now imported from from metadata.ingestion.api.steps import Source (instead of from metadata.ingestion.api.source import Source)
  3. We are now initializing the OpenMetadata object at the Workflow level (to share it better in each step). Therefore, the source __init__ method signature is now def __init__(self, config: WorkflowSource, metadata: OpenMetadata):. Make sure to store the self.metadata object during the __init__ and don't forget to call super().__init__().
  4. We are updating how the status & exception management happens in the connectors. Now each yield result is wrapped by an Either (imported from from metadata.ingestion.api.models import Either). Your correct data will be yielded in a right, while the errors are tracked in a left. Read more about the Workflow management here.
  • Pipeline Status are now timestamps in milliseconds.

Upgrade Process

You can get changes from artifact hub of openmetadata helm chart release. Click on Default Values >> Compare to Version.

Helm Chart Release Comparison

Update Helm Chart Locally for OpenMetadata with the below command:

It will result in the below output on screen.

Verify with the below command to see the latest release available locally.

You can run the below command to upgrade the dependencies with the new chart

The above command uses configurations defined here. You can modify any configuration and deploy by passing your own values.yaml.

Make sure that, when using your own values.yaml, you are not overwriting elements such as the image of the containers. This would prevent your new deployment to use the latest containers when running the upgrade.

If you are running into any issues, double-check what are the default values of the helm revision.

Finally, we upgrade OpenMetadata with the below command:

You might need to pass your own values.yaml with the --values flag.

Note that in every version upgrade there is a migration process that updates your database to the newest version.

For kubernetes, this process will happen automatically as an upgrade hook.

You can learn more about how the migration process works here.

Post-Upgrade Steps

Go to Settings -> Applications -> Search Indexing



Click on Run Now.

In the configuration section, you can select the entities you want to reindex.



Since this is required after the upgrade, we want to reindex All the entities.

If you are running the ingestion workflows externally or using a custom Airflow installation, you need to make sure that the Python Client you use is aligned with the OpenMetadata server version.

For example, if you are upgrading the server to the version x.y.z, you will need to update your client with

The plugin parameter is a list of the sources that we want to ingest. An example would look like this openmetadata-ingestion[mysql,snowflake,s3]==1.2.0. You will find specific instructions for each connector here.

Moreover, if working with your own Airflow deployment - not the openmetadata-ingestion image - you will need to upgrade as well the openmetadata-managed-apis version:

Go to Settings -> {service entity} -> Pipelines



Select the pipelines you want to Re Deploy click Re Deploy.

If you are seeing broken dags select all the pipelines from all the services and re deploy the pipelines.


With Release 1.0.0, if you see your helm charts failing to deploy with the below issue -

This means the values passed to the helm charts has a section global.airflow. As per the breaking changes mentioned here, Airflow configs are replaced with pipelineServiceClient for Helm Charts.

The Helm Chart Values JSON Schema helps to catch the above breaking changes and this section will help you resolve and update your configurations for the same. You can read more about JSON Schema with Helm Charts here.

You will need to update the existing section of global.airflow values to match the new configurations.

⛔ Before 1.0.0 Helm Chart Release, the global.airflow section would be like -

✅ After 1.0.0 Helm Chart Release, the global.pipelineServiceClient section will replace the above airflow section -

Run the helm lint command on your custom values after making the changes to validate with the JSON Schema.

This issue will only occur if you are using openmetadata-dependencies helm chart version 0.0.49 and 0.0.50 and upgrading to latest helm chart release.

If your helm dependencies upgrade fails with the below command result -

This issue is related to a minor change that affected the MySQL Database Engine version upgrade from 8.0.28 to 8.0.29 for the Helm Chart Release 0.0.49 and 0.0.50. Then the registry url was updated as we found a work around to fetch previous versions of bitnami/mysql Helm Releases.

As a result of the above fixes, anyone who is on OpenMetadata Dependencies Helm Chart Version 0.0.49 and 0.0.50 is affected with the above issue when upgrading for mysql. In order to fix this issue, make sure to follow the below steps -

  1. Backup the Database using Metadata Backup CLI as mentioned here
  2. Uninstall OpenMetadata Dependencies Helm Chart (helm uninstall openmetadata-dependencies)
  3. Remove the unmanaged volume for MySQL Stateful Set Kubernetes Object (kubectl delete pvc data-mysql-0)
  4. Install the latest version of OpenMetadata Dependencies Helm Chart
  5. Restore the Database using Metadata Restore CLI as mentioned here
  6. Next, Proceed with upgrade for OpenMetadata Helm Chart as mentioned here