kafka batch scram source Kafka Batch Scram Source

Provided by: "Apache Software Foundation"

Support Level for this Kamelet is: "Preview"

Receive data from Kafka topics in batch through SCRAM login module and commit them manually through KafkaManualCommit..

Configuration Options

The following table summarizes the configuration options available for the kafka-batch-scram-source Kamelet:

Property Name Description Type Default Example

bootstrapServers

Bootstrap Servers

Required Comma separated list of Kafka Broker URLs.

string

password

Password

Required Password to authenticate to kafka.

string

topic

Topic Names

Required Comma separated list of Kafka topic names.

string

user

Username

Required Username to authenticate to Kafka.

string

allowManualCommit

Allow Manual Commit

Whether to allow doing manual commits.

boolean

false

autoCommitEnable

Auto Commit Enable

If true, periodically commit to ZooKeeper the offset of messages already fetched by the consumer.

boolean

true

autoOffsetReset

Auto Offset Reset

What to do when there is no initial offset. There are 3 enums and the value can be one of latest, earliest, none.

string

latest

batchSize

Batch Dimension

The maximum number of records returned in a single call to poll().

int

500

consumerGroup

Consumer Group

A string that uniquely identifies the group of consumers to which this source belongs.

string

my-group-id

deserializeHeaders

Automatically Deserialize Headers

When enabled the Kamelet source will deserialize all message headers to String representation.

boolean

true

maxPollIntervalMs

Max Poll Interval

The maximum delay between invocations of poll() when using consumer group management.

int

pollOnError

Poll On Error Behavior

What to do if kafka threw an exception while polling for new messages. There are 5 enums and the value can be one of DISCARD, ERROR_HANDLER, RECONNECT, RETRY, STOP.

string

ERROR_HANDLER

pollTimeout

Poll Timeout Interval

The timeout used when polling the KafkaConsumer.

int

5000

saslMechanism

SASL Mechanism

The Simple Authentication and Security Layer (SASL) Mechanism used.

string

SCRAM-SHA-512

securityProtocol

Security Protocol

Protocol used to communicate with brokers. SASL_PLAINTEXT, PLAINTEXT, SASL_SSL and SSL are supported.

string

SASL_SSL

Dependencies

At runtime, the kafka-batch-scram-source Kamelet relies upon the presence of the following dependencies:

  • mvn:org.apache.camel.kamelets:camel-kamelets-utils:4.4.4-SNAPSHOT

  • camel:core

  • camel:kafka

  • camel:kamelet

Camel JBang usage

Prerequisites

  • You’ve installed JBang.

  • You have executed the following command:

jbang app install camel@apache/camel

Supposing you have a file named route.yaml with this content:

- route:
    from:
      uri: "kamelet:timer-source"
      parameters:
        period: 10000
        message: 'test'
      steps:
        - to:
            uri: "kamelet:log-sink"

You can now run it directly through the following command

camel run route.yaml

Camel K Environment Usage

This section describes how you can use the kafka-batch-scram-source.

Knative source

You can use the kafka-batch-scram-source Kamelet as a Knative source by binding it to a Knative object.

kafka-batch-scram-source-pipe.yaml
apiVersion: camel.apache.org/v1
kind: Pipe
metadata:
  name: kafka-batch-scram-source-pipe
spec:
  source:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1
      name: kafka-batch-scram-source
    properties:
      bootstrapServers: The Bootstrap Servers
      password: The Password
      topic: The Topic Names
      user: The Username
  sink:
    ref:
      kind: Channel
      apiVersion: messaging.knative.dev/v1
      name: mychannel

Prerequisite

You have Camel K installed on the cluster.

Procedure for using the cluster CLI

  1. Save the kafka-batch-scram-source-pipe.yaml file to your local drive, and then edit it as needed for your configuration.

  2. Run the source by using the following command:

    kubectl apply -f kafka-batch-scram-source-pipe.yaml

Procedure for using the Kamel CLI

Configure and run the source by using the following command:

kamel bind channel:mychannel -p "source.bootstrapServers=The Bootstrap Servers" -p "source.password=The Password" -p "source.topic=The Topic Names" -p "source.user=The Username" kafka-batch-scram-source

This command creates the Kamelet Pipe in the current namespace on the cluster.

Kafka source

You can use the kafka-batch-scram-source Kamelet as a Kafka source by binding it to a Kafka topic.

kafka-batch-scram-source-pipe.yaml
apiVersion: camel.apache.org/v1
kind: Pipe
metadata:
  name: kafka-batch-scram-source-pipe
spec:
  source:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1
      name: kafka-batch-scram-source
    properties:
      bootstrapServers: The Bootstrap Servers
      password: The Password
      topic: The Topic Names
      user: The Username
  sink:
    ref:
      kind: KafkaTopic
      apiVersion: kafka.strimzi.io/v1beta1
      name: my-topic

Prerequisites

  • You’ve installed Strimzi.

  • You’ve created a topic named my-topic in the current namespace.

  • You have Camel K installed on the cluster.

Procedure for using the cluster CLI

  1. Save the kafka-batch-scram-source-pipe.yaml file to your local drive, and then edit it as needed for your configuration.

  2. Run the source by using the following command:

    kubectl apply -f kafka-batch-scram-source-pipe.yaml

Procedure for using the Kamel CLI

Configure and run the source by using the following command:

kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic -p "source.bootstrapServers=The Bootstrap Servers" -p "source.password=The Password" -p "source.topic=The Topic Names" -p "source.user=The Username" kafka-batch-scram-source

This command creates the Kamelet Pipe in the current namespace on the cluster.