Kafka
This page guides you through the process of setting up the Kafka source connector.
Set up guide
Step 1: Set up Kafka
To use the Kafka source connector, you'll need:
- A Kafka cluster 1.0 or above
- Airbyte user should be allowed to read messages from topics, and these topics should be created before reading from Kafka.
Step 2: Setup the Kafka source in Airbyte
You'll need the following information to configure the Kafka source:
- Group ID - The Group ID is how you distinguish different consumer groups. (e.g. group.id)
- Protocol - The Protocol used to communicate with brokers.
- Client ID - An ID string to pass to the server when making requests. The purpose of this is to be able to track the source of requests beyond just ip/port by allowing a logical application name to be included in server-side request logging. (e.g. airbyte-consumer)
- Test Topic - The Topic to test in case the Airbyte can consume messages. (e.g. test.topic)
- Subscription Method - You can choose to manually assign a list of partitions, or subscribe to all topics matching specified pattern to get dynamically assigned partitions.
- List of topic
- Bootstrap Servers - A list of host/port pairs to use for establishing the initial connection to the Kafka cluster.
- Schema Registry - Host/port to connect schema registry server. Note: It supports for AVRO format only.
For Airbyte Open Source:
- Go to the Airbyte UI and in the left navigation bar, click Sources. In the top-right corner, click +new source.
- On the Set up the source page, enter the name for the Kafka connector and select Kafka from the Source type dropdown.
- Follow the Setup the Kafka source in Airbyte
Supported sync modes
The Kafka source connector supports the following sync modes:
Feature | Supported?(Yes/No) | Notes |
---|---|---|
Full Refresh Sync | Yes | |
Incremental - Append Sync | Yes | |
Namespaces | No |
Supported Format
JSON - Json value messages. It does not support schema registry now.
AVRO - deserialize Using confluent API. Please refer (https://docs.confluent.io/platform/current/schema-registry/serdes-develop/serdes-avro.html)
Changelog
Version | Date | Pull Request | Subject |
---|---|---|---|
0.2.4 | 2024-02-13 | 35229 | Adopt CDK 0.20.4 |
0.2.4 | 2024-01-24 | 34453 | bump CDK version |
0.2.3 | 2022-12-06 | 19587 | Fix missing data before consumer is closed |
0.2.2 | 2022-11-04 | 18648 | Add missing record_count increment for JSON |
0.2.1 | 2022-11-04 | This version was the same as 0.2.0 and was committed so using 0.2.2 next to keep versions in order | |
0.2.0 | 2022-08-22 | 13864 | Added AVRO format support and Support for maximum records to process |
0.1.7 | 2022-06-17 | 13864 | Updated stacktrace format for any trace message errors |
0.1.6 | 2022-05-29 | 12903 | Add Polling Time to Specification (default 100 ms) |
0.1.5 | 2022-04-19 | 12134 | Add PLAIN Auth |
0.1.4 | 2022-02-15 | 10186 | Add SCRAM-SHA-512 Auth |
0.1.3 | 2022-02-14 | 10256 | Add -XX:+ExitOnOutOfMemoryError JVM option |
0.1.2 | 2021-12-21 | 8865 | Fix SASL config read issue |
0.1.1 | 2021-12-06 | 8524 | Update connector fields title/description |