Upgrade on Bare Metal
This guide will help you upgrade an OpenMetadata deployment using release binaries.
Requirements
This guide assumes that you have an OpenMetadata deployment that you installed and configured following the Bare Metal deployment guide.
Prerequisites
Everytime that you plan on upgrading OpenMetadata to a newer version, make sure to go over all these steps:
Backup your Metadata
Before upgrading your OpenMetadata version we strongly recommend backing up the metadata.
The source of truth is stored in the underlying database (MySQL and Postgres supported). During each version upgrade there is a database migration process that needs to run. It will directly attack your database and update the shape of the data to the newest OpenMetadata release.
It is important that we backup the data because if we face any unexpected issues during the upgrade process, you will be able to get back to the previous version without any loss.
You can learn more about how the migration process works here.
During the upgrade, please note that the backup is only for safety and should not be used to restore data to a higher version.
Since version 1.4.0, OpenMetadata encourages using the builtin-tools for creating logical backups of the metadata:
For PROD deployment we recommend users to rely on cloud services for their databases, be it AWS RDS, Azure SQL or GCP Cloud SQL.
If you're a user of these services, you can leverage their backup capabilities directly:
You can refer to the following guide to get more details about the backup and restore:
Understanding the "Running" State in OpenMetadata
In OpenMetadata, the "Running" state indicates that the OpenMetadata server has received a response from Airflow confirming that a workflow is in progress. However, if Airflow unexpectedly stops or crashes before it can send a failure status update through the Failure Callback, OpenMetadata remains unaware of the workflow’s actual state. As a result, the workflow may appear to be stuck in "Running" even though it is no longer executing.
This situation can also occur during an OpenMetadata upgrade. If an ingestion pipeline was running at the time of the upgrade and the process caused Airflow to shut down, OpenMetadata would not receive any further updates from Airflow. Consequently, the pipeline status remains "Running" indefinitely.

Running State in OpenMetadata
Expected Steps to Resolve
To resolve this issue:
- Ensure that Airflow is restarted properly after an unexpected shutdown.
- Manually update the pipeline status if necessary.
- Check Airflow logs to verify if the DAG execution was interrupted.
Update sort_buffer_size
(MySQL) or work_mem
(Postgres)
Before running the migrations, it is important to update these parameters to ensure there are no runtime errors. A safe value would be setting them to 20MB.
If using MySQL
You can update it via SQL (note that it will reset after the server restarts):
To make the configuration persistent, you'd need to navigate to your MySQL Server install directory and update the my.ini
or my.cnf
files with sort_buffer_size = 20971520
.
If using RDS, you will need to update your instance's Parameter Group to include the above change.
If using Postgres
You can update it via SQL (not that it will reset after the server restarts):
To make the configuration persistent, you'll need to update the postgresql.conf
file with work_mem = 20MB
.
If using RDS, you will need to update your instance's Parameter Group to include the above change.
Note that this value would depend on the size of your query_entity
table. If you still see Out of Sort Memory Error
s during the migration after bumping this value, you can increase them further.
After the migration is finished, you can revert this changes.
Backward Incompatible Changes
1.7.0
Removing support for Python 3.8
Python 3.8 was officially EOL on 2024-10-07. Some of our dependencies have already started removing support for higher versions, and are following suit to ensure we are using the latest and most stable versions of our dependencies.
This means that for Release 1.7, the supported Python versions for the Ingestion Framework are 3.9, 3.10 and 3.11.
We were already shipping our Docker images with Python 3.10, so this change should not affect you if you are using our Docker images. However, if you installed the openmetadata-ingestion
package directly, please make sure to update your Python version to 3.9 or higher.
OpenSearch Settings Update
OpenSearch has a different default value of max_clause_count
than Elasticsearch. This means that if you are using OpenSearch, you will need to update the max_clause_count
setting in your OpenSearch configuration:
If you're using AWS, you can add this setting from the console as well

AWS OpenSearch Settings
Upgrade process
Step 1: Download the binaries for the release you want to install
OpenMetadata release binaries are maintained as GitHub releases.
To download a specific release binary:
- Visit github.com/open-metadata/OpenMetadata/releases. The latest release will be at the top of this page.
- Locate the Assets' section for the release you want to upgrade to.
- Download the release binaries. The release binaries will be in a compressed tar file named using the following convention,
openmetadata-x.y.z.tar.gz
Wherex
,y
,z
are the major, minor, and patch release numbers, respectively.
Step 2: Extract the release binaries from the download file
Using the command-line tool or application of your choice, extract the release binaries.
For example, to extract using tar
, run the following command.
This will create a directory with the same name as the download file minus the .tar
and .gz
extensions.
Step 3: Navigate into the directory created by extracting the release binaries
Change into the new directory by issuing a command similar to the following.
For example, to navigate into the directory created by issuing the tar command above, you would run the following command.
Step 4: Stop the OpenMetadata server
OpenMetadata ships with a few control scripts. One is openmetadata.sh
. This script enables you to start, stop, and perform other deployment operations on the OpenMetadata server.
Most OpenMetadata releases will require you to migrate your data to updated schemas.
Before you migrate your data to the new release you are upgrading to, stop the OpenMetadata server from the directory of your current installation by running the following command:
Step 5: Migrate the database schemas and ElasticSearch indexes
The bootstrap/openmetadata-ops.sh
script enables you to perform a number of operations on the OpenMetadata database (in MySQL) and index (in Elasticsearch).
Step 6: Restart the OpenMetadata server
Once you've dropped and recreated your data in the new version, restart the OpenMetadata server using the new release binaries. You may restart the server by running the following command.
Post-Upgrade Steps
Reindex
With UI
Since this is required after the upgrade, we want to reindex All
the entities.
(Optional) Update your OpenMetadata Ingestion Client
If you are running the ingestion workflows externally or using a custom Airflow installation, you need to make sure that the Python Client you use is aligned with the OpenMetadata server version.
For example, if you are upgrading the server to the version x.y.z
, you will need to update your client with
With Kubernetes
Follow these steps to reindex using the CLI:
- List the CronJobs Use the following command to check the available CronJobs:
Upon running this command you should see output similar to the following.
- Create a Job from a CronJob Create a one-time job from an existing CronJob using the following command:
Replace <job_name>
with the actual name of the job.
Upon running this command you should see output similar to the following.
- Check the Job Status Verify the status of the created job with:
Upon running this command you should see output similar to the following.
- view logs To view the logs use the below command.
Replace <job_name>
with the actual job name.
The plugin
parameter is a list of the sources that we want to ingest. An example would look like this openmetadata-ingestion[mysql,snowflake,s3]==1.2.0
. You will find specific instructions for each connector here.
Moreover, if working with your own Airflow deployment - not the openmetadata-ingestion
image - you will need to upgrade as well the openmetadata-managed-apis
version:
Re Deploy Ingestion Pipelines
With UI
With Kubernetes
Follow these steps to deploy pipelines using the CLI:
- List the CronJobs Use the following command to check the available CronJobs:
Upon running this command you should see output similar to the following.
- Create a Job from a CronJob Create a one-time job from an existing CronJob using the following command:
Replace <job_name>
with the actual name of the job.
Upon running this command you should see output similar to the following.
- Check the Job Status Verify the status of the created job with:
Upon running this command you should see output similar to the following.
- view logs To view the logs use the below command.
Replace <job_name>
with the actual job name.
If you are seeing broken dags select all the pipelines from all the services and re deploy the pipelines.
Openmetadata-ops Script
Overview
The openmetadata-ops
script is designed to manage and migrate databases and search indexes, reindex existing data into Elastic Search or OpenSearch, and redeploy service pipelines.
Usage
Commands
- analyze-tables
Migrates secrets from the database to the configured Secrets Manager. Note that this command does not support migrating between external Secrets Managers.
- changelog
Prints the change log of database migration.
- check-connection
Checks if a connection can be successfully obtained for the target database.
- deploy-pipelines
Deploys all the service pipelines.
- drop-create
Deletes any tables in the configured database and creates new tables based on the current version of OpenMetadata. This command also re-creates the search indexes.
- info
Shows the list of migrations applied and the pending migrations waiting to be applied on the target database.
- migrate
Migrates the OpenMetadata database schema and search index mappings.
- migrate-secrets
Migrates secrets from the database to the configured Secrets Manager. Note that this command does not support migrating between external Secrets Managers.
- reindex
Reindexes data into the search engine from the command line.
- repair
Repairs the DATABASE_CHANGE_LOG table, which is used to track all the migrations on the target database. This involves removing entries for the failed migrations and updating the checksum of migrations already applied on the target database.
- validate
Checks if all the migrations have been applied on the target database.
Examples
Display Help To display the help message:
Migrate Database Schema
To migrate the database schema and search index mappings:
Reindex Data
To reindex data into the search engine: