From d4bf2b7400df95df6c3f2534519671e596163be2 Mon Sep 17 00:00:00 2001 From: Kai Moritz Date: Tue, 14 Dec 2021 18:55:02 +0100 Subject: [PATCH] First Contact: Simple Producer & Consumer --- .gitignore | 3 + README.sh | 33 +++++ docker-compose.yml | 37 +++++ pom.xml | 58 ++++++++ .../java/de/juplo/kafka/SimpleConsumer.java | 131 ++++++++++++++++++ .../java/de/juplo/kafka/SimpleProducer.java | 88 ++++++++++++ src/main/resources/logback.xml | 17 +++ 7 files changed, 367 insertions(+) create mode 100644 .gitignore create mode 100755 README.sh create mode 100644 docker-compose.yml create mode 100644 pom.xml create mode 100644 src/main/java/de/juplo/kafka/SimpleConsumer.java create mode 100644 src/main/java/de/juplo/kafka/SimpleProducer.java create mode 100644 src/main/resources/logback.xml diff --git a/.gitignore b/.gitignore new file mode 100644 index 0000000..6240411 --- /dev/null +++ b/.gitignore @@ -0,0 +1,3 @@ +*.iml +.idea +target diff --git a/README.sh b/README.sh new file mode 100755 index 0000000..95aae0b --- /dev/null +++ b/README.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +if [ "$1" = "cleanup" ] +then + docker-compose down -v + mvn clean + exit +fi + +mvn package || exit 1 +if [ "$1" = "build" ]; then exit; fi + +trap 'kill $(jobs -p) 2>/dev/null' EXIT + +docker-compose up -d + +echo "Waiting for the Kafka-Cluster to become ready..." +docker-compose exec kafka cub kafka-ready -b kafka:9092 1 60 > /dev/null 2>&1 || exit 1 + +echo "Producing messages" +mvn exec:java@producer + +echo "Reading messages" +mvn exec:java@consumer & +sleep 7 +kill $(jobs -p) +sleep 2 + +echo "Re-Reading messages" +mvn exec:java@consumer & +sleep 7 +kill $(jobs -p) +sleep 2 diff --git a/docker-compose.yml b/docker-compose.yml new file mode 100644 index 0000000..13e950d --- /dev/null +++ b/docker-compose.yml @@ -0,0 +1,37 @@ +version: '3.2' +services: + zookeeper: + image: confluentinc/cp-zookeeper:6.2.0 + environment: + ZOOKEEPER_CLIENT_PORT: 2181 + ports: + - 2181:2181 + + kafka: + image: confluentinc/cp-kafka:6.2.0 + environment: + KAFKA_BROKER_ID: 1 + KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181 + KAFKA_LISTENERS: DOCKER://:9092, LOCALHOST://:9082 + KAFKA_ADVERTISED_LISTENERS: DOCKER://kafka:9092, LOCALHOST://localhost:9082 + KAFKA_INTER_BROKER_LISTENER_NAME: DOCKER + KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: DOCKER:PLAINTEXT, LOCALHOST:PLAINTEXT + KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1 + KAFKA_AUTO_CREATE_TOPICS_ENABLE: "false" + ports: + - 9092:9082 + - 9082:9082 + depends_on: + - zookeeper + + setup: + image: juplo/toolbox + command: > + bash -c " + kafka-topics --bootstrap-server kafka:9092 --delete --if-exists --topic test + kafka-topics --bootstrap-server kafka:9092 --create --topic test --partitions 2 + " + + cli: + image: juplo/toolbox + command: sleep infinity diff --git a/pom.xml b/pom.xml new file mode 100644 index 0000000..39a9e4e --- /dev/null +++ b/pom.xml @@ -0,0 +1,58 @@ + + + + 4.0.0 + + + org.springframework.boot + spring-boot-starter-parent + 2.6.0 + + + + de.juplo.kafka + first-contact + First Contact: a Simple Producer and a simple Consumer-Group + 1.0-SNAPSHOT + + + + org.apache.kafka + kafka-clients + + + org.projectlombok + lombok + + + ch.qos.logback + logback-classic + + + + + + + org.codehaus.mojo + exec-maven-plugin + 3.0.0 + + + producer + + de.juplo.kafka.SimpleProducer + + + + consumer + + de.juplo.kafka.SimpleConsumer + + + + + + + + diff --git a/src/main/java/de/juplo/kafka/SimpleConsumer.java b/src/main/java/de/juplo/kafka/SimpleConsumer.java new file mode 100644 index 0000000..e4d9697 --- /dev/null +++ b/src/main/java/de/juplo/kafka/SimpleConsumer.java @@ -0,0 +1,131 @@ +package de.juplo.kafka; + +import lombok.extern.slf4j.Slf4j; +import org.apache.kafka.clients.consumer.ConsumerRecord; +import org.apache.kafka.clients.consumer.ConsumerRecords; +import org.apache.kafka.clients.consumer.KafkaConsumer; +import org.apache.kafka.common.errors.WakeupException; +import org.apache.kafka.common.serialization.StringDeserializer; + +import java.time.Duration; +import java.util.Arrays; +import java.util.Properties; +import java.util.concurrent.locks.Condition; +import java.util.concurrent.locks.Lock; +import java.util.concurrent.locks.ReentrantLock; + + +@Slf4j +public class SimpleConsumer +{ + private long consumed = 0; + private KafkaConsumer consumer; + private Lock lock = new ReentrantLock(); + private Condition stopped = lock.newCondition(); + + + public SimpleConsumer() + { + // tag::create[] + Properties props = new Properties(); + props.put("bootstrap.servers", ":9092"); + props.put("group.id", "my-consumer"); // << Used for Offset-Commits + // end::create[] + props.put("auto.offset.reset", "earliest"); + // tag::create[] + props.put("key.deserializer", StringDeserializer.class.getName()); + props.put("value.deserializer", StringDeserializer.class.getName()); + + KafkaConsumer consumer = new KafkaConsumer<>(props); + // end::create[] + this.consumer = consumer; + } + + + public void run() + { + String id = "C"; + + try + { + log.info("{} - Subscribing to topic test", id); + consumer.subscribe(Arrays.asList("test")); + + // tag::loop[] + while (true) + { + ConsumerRecords records = + consumer.poll(Duration.ofSeconds(1)); + + // Do something with the data... + // end::loop[] + log.info("{} - Received {} messages", id, records.count()); + for (ConsumerRecord record : records) + { + consumed++; + log.info( + "{} - {}: {}/{} - {}={}", + id, + record.offset(), + record.topic(), + record.partition(), + record.key(), + record.value() + ); + } + // tag::loop[] + } + // end::loop[] + } + catch(WakeupException e) + { + log.info("{} - RIIING!", id); + } + catch(Exception e) + { + log.error("{} - Unexpected error: {}", id, e.toString()); + } + finally + { + this.lock.lock(); + try + { + log.info("{} - Closing the KafkaConsumer", id); + consumer.close(); + log.info("C - DONE!"); + stopped.signal(); + } + finally + { + this.lock.unlock(); + log.info("{}: Consumed {} messages in total, exiting!", id, consumed); + } + } + } + + + public static void main(String[] args) throws Exception + { + SimpleConsumer instance = new SimpleConsumer(); + + Runtime.getRuntime().addShutdownHook(new Thread(() -> + { + instance.lock.lock(); + try + { + instance.consumer.wakeup(); + instance.stopped.await(); + } + catch (InterruptedException e) + { + log.warn("Interrrupted while waiting for the consumer to stop!", e); + } + finally + { + instance.lock.unlock(); + } + })); + + instance.run(); + } +} diff --git a/src/main/java/de/juplo/kafka/SimpleProducer.java b/src/main/java/de/juplo/kafka/SimpleProducer.java new file mode 100644 index 0000000..5f57925 --- /dev/null +++ b/src/main/java/de/juplo/kafka/SimpleProducer.java @@ -0,0 +1,88 @@ +package de.juplo.kafka; + +import lombok.extern.slf4j.Slf4j; +import org.apache.kafka.clients.producer.KafkaProducer; +import org.apache.kafka.clients.producer.ProducerRecord; +import org.apache.kafka.common.serialization.StringSerializer; + +import java.util.Properties; + + +@Slf4j +public class SimpleProducer +{ + public static void main(String[] args) throws Exception + { + // tag::create[] + Properties props = new Properties(); + props.put("bootstrap.servers", "localhost:9092"); + props.put("key.serializer", StringSerializer.class.getName()); + props.put("value.serializer", StringSerializer.class.getName()); + + KafkaProducer producer = new KafkaProducer<>(props); + // end::create[] + + String id = "P"; + long i = 0; + + try + { + for (; i < 100 ; i++) + { + final long time = System.currentTimeMillis(); + + final ProducerRecord record = new ProducerRecord<>( + "test", // Topic + Long.toString(i%10), // Key + Long.toString(i) // Value + ); + + producer.send(record, (metadata, e) -> + { + long now = System.currentTimeMillis(); + if (e == null) + { + // HANDLE SUCCESS + log.debug( + "{} - Sent key={} message={} partition={}/{} timestamp={} latency={}ms", + id, + record.key(), + record.value(), + metadata.partition(), + metadata.offset(), + metadata.timestamp(), + now - time + ); + } + else + { + // HANDLE ERROR + log.error( + "{} - ERROR key={} timestamp={} latency={}ms: {}", + id, + record.key(), + metadata == null ? -1 : metadata.timestamp(), + now - time, + e.toString() + ); + } + }); + + long now = System.currentTimeMillis(); + log.trace( + "{} - Queued #{} key={} latency={}ms", + id, + i, + record.key(), + now - time + ); + } + } + finally + { + log.info("{}: Closing the KafkaProducer", id); + producer.close(); + log.info("{}: Exiting!", id); + } + } +} diff --git a/src/main/resources/logback.xml b/src/main/resources/logback.xml new file mode 100644 index 0000000..b8e6780 --- /dev/null +++ b/src/main/resources/logback.xml @@ -0,0 +1,17 @@ + + + + + + %highlight(%-5level) %m%n + + + + + + + + + + + -- 2.20.1