kafka sink Kafka Sink

Provided by: "Apache Software Foundation"

Support Level for this Kamelet is: "Preview"

Send data to Kafka topics.

The Kamelet is able to understand the following headers to be set:

  • key / ce-key: as message key

  • partition-key / ce-partitionkey: as message partition key

Both the headers are optional.

Configuration Options

The following table summarizes the configuration options available for the kafka-sink Kamelet:

Property Name Description Type Default Example

bootstrapServers

Brokers

Required Comma separated list of Kafka Broker URLs.

string

password

Password

Required Password to authenticate to kafka.

string

topic

Topic Names

Required Comma separated list of Kafka topic names.

string

user

Username

Required Username to authenticate to Kafka.

string

saslMechanism

SASL Mechanism

The Simple Authentication and Security Layer (SASL) Mechanism used.

string

PLAIN

securityProtocol

Security Protocol

Protocol used to communicate with brokers. SASL_PLAINTEXT, PLAINTEXT, SASL_SSL and SSL are supported.

string

SASL_SSL

Dependencies

At runtime, the kafka-sink Kamelet relies upon the presence of the following dependencies:

  • camel:kafka

  • camel:kamelet

Usage

This section describes how you can use the kafka-sink.

Knative sink

You can use the kafka-sink Kamelet as a Knative sink by binding it to a Knative object.

kafka-sink-binding.yaml
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: kafka-sink-binding
spec:
  source:
    ref:
      kind: Channel
      apiVersion: messaging.knative.dev/v1
      name: mychannel
  sink:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: kafka-sink
    properties:
      bootstrapServers: The Brokers
      password: The Password
      topic: The Topic Names
      user: The Username

Prerequisite

You have Camel K installed on the cluster.

Procedure for using the cluster CLI

  1. Save the kafka-sink-binding.yaml file to your local drive, and then edit it as needed for your configuration.

  2. Run the sink by using the following command:

    kubectl apply -f kafka-sink-binding.yaml

Procedure for using the Kamel CLI

Configure and run the sink by using the following command:

kamel bind kafka-sink -p "sink.bootstrapServers=The Brokers" -p "sink.password=The Password" -p "sink.topic=The Topic Names" -p "sink.user=The Username" channel:mychannel

This command creates the KameletBinding in the current namespace on the cluster.

Kafka sink

You can use the kafka-sink Kamelet as a Kafka sink by binding it to a Kafka topic.

kafka-sink-binding.yaml
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: kafka-sink-binding
spec:
  source:
    ref:
      kind: KafkaTopic
      apiVersion: kafka.strimzi.io/v1beta1
      name: my-topic
  sink:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: kafka-sink
    properties:
      bootstrapServers: The Brokers
      password: The Password
      topic: The Topic Names
      user: The Username

Prerequisites

  • You’ve installed Strimzi.

  • You’ve created a topic named my-topic in the current namespace.

  • You have Camel K installed on the cluster.

Procedure for using the cluster CLI

  1. Save the kafka-sink-binding.yaml file to your local drive, and then edit it as needed for your configuration.

  2. Run the sink by using the following command:

    kubectl apply -f kafka-sink-binding.yaml

Procedure for using the Kamel CLI

Configure and run the sink by using the following command:

kamel bind kafka-sink -p "sink.bootstrapServers=The Brokers" -p "sink.password=The Password" -p "sink.topic=The Topic Names" -p "sink.user=The Username" kafka.strimzi.io/v1beta1:KafkaTopic:my-topic

This command creates the KameletBinding in the current namespace on the cluster.