
KafkaConnect
PRODFeature List
✓ Pipelines
✓ Pipeline Status
✓ Tags
✓ Usage
✕ Owners
✕ Lineage
How to Run the Connector Externally
To run the Ingestion via the UI you’ll need to use the OpenMetadata Ingestion Container, which comes shipped with custom Airflow plugins to handle the workflow deployment. If, instead, you want to manage your workflows externally on your preferred orchestrator, you can check the following docs to run the Ingestion Framework anywhere.Requirements
Python Requirements
To run the KafkaConnect ingestion, you will need to install:Metadata Ingestion
All connectors are defined as JSON Schemas. Here you can find the structure to create a connection to KafkaConnect. In order to create and run a Metadata Ingestion workflow, we will follow the steps to create a YAML configuration able to connect to the source, process the Entities if needed, and reach the OpenMetadata server. The workflow is modeled around the following JSON Schema1. Define the YAML Config
This is a sample config for KafkaConnect:2. Run with the CLI
First, we will need to save the YAML file. Afterward, and with all requirements installed, we can run:Debezium CDC Support
The KafkaConnect connector provides full support for Debezium CDC connectors with intelligent column extraction and accurate lineage tracking.What We Provide
When you ingest Debezium connectors, OpenMetadata automatically:- Detects CDC Envelope Structures - Identifies Debezium’s CDC format with
op,before, andafterfields - Extracts Real Table Columns - Parses actual database columns from the CDC payload instead of CDC envelope metadata
- Creates Accurate Column-Level Lineage - Maps lineage from source database tables → Kafka topics → target systems
Recognized Configuration Parameters
OpenMetadata recognizes the following Debezium configuration parameters for intelligent CDC detection:database.server.name- Server identifier (Debezium V1)topic.prefix- Topic prefix (Debezium V2)table.include.list- Tables to capture (e.g.,mydb.customers,mydb.orders)