From 0ca17e39b4f1fe09482d876b2cb780ebeaead059 Mon Sep 17 00:00:00 2001 From: Kai Moritz Date: Thu, 26 Sep 2024 14:57:20 +0200 Subject: [PATCH] `spring-producer` zu einem `rest-producer` erweitert --- README.sh | 98 +++++++++++++++++-- docker/docker-compose.yml | 39 +++++++- pom.xml | 6 +- src/main/java/de/juplo/kafka/Application.java | 3 + .../juplo/kafka/ApplicationConfiguration.java | 1 + .../de/juplo/kafka/ApplicationProperties.java | 1 + .../java/de/juplo/kafka/ErrorResponse.java | 11 +++ .../java/de/juplo/kafka/ExampleProducer.java | 77 ++++++--------- .../java/de/juplo/kafka/ProduceFailure.java | 21 ++++ .../java/de/juplo/kafka/ProduceResult.java | 11 +++ .../java/de/juplo/kafka/ProduceSuccess.java | 12 +++ .../java/de/juplo/kafka/ApplicationTests.java | 10 +- 12 files changed, 229 insertions(+), 61 deletions(-) create mode 100644 src/main/java/de/juplo/kafka/ErrorResponse.java create mode 100644 src/main/java/de/juplo/kafka/ProduceFailure.java create mode 100644 src/main/java/de/juplo/kafka/ProduceResult.java create mode 100644 src/main/java/de/juplo/kafka/ProduceSuccess.java diff --git a/README.sh b/README.sh index 9148486..d23ed67 100755 --- a/README.sh +++ b/README.sh @@ -1,6 +1,6 @@ #!/bin/bash -IMAGE=juplo/spring-producer:1.0-SNAPSHOT +IMAGE=juplo/rest-producer:1.0-SNAPSHOT if [ "$1" = "cleanup" ] then @@ -10,7 +10,7 @@ then fi docker compose -f docker/docker-compose.yml up -d --remove-orphans kafka-1 kafka-2 kafka-3 -docker compose -f docker/docker-compose.yml rm -svf producer +docker compose -f docker/docker-compose.yml rm -svf producer producer-0 producer-1 if [[ $(docker image ls -q $IMAGE) == "" || @@ -29,12 +29,94 @@ docker compose -f docker/docker-compose.yml up -t0 -d cli sleep 1 docker compose -f docker/docker-compose.yml logs setup -docker compose -f docker/docker-compose.yml ps docker compose -f docker/docker-compose.yml up -d producer -sleep 5 +while ! [[ $(http 0:8080/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer..."; sleep 1; done -docker compose -f docker/docker-compose.yml exec cli kafkacat -b kafka:9092 -t test -c 20 -f'topic=%t\tpartition=%p\toffset=%o\tkey=%k\tvalue=%s\n' +# tag::hashed[] +echo -n Nachricht 1 an klaus über producer | http -v :8080/klaus +# end::hashed[] +echo -n Nachricht 2 an klaus über producer | http -v :8080/klaus +# tag::hashed[] +echo -n Nachricht 1 an peter über producer | http -v :8080/peter +# end::hashed[] +echo -n Nachricht 3 an klaus über producer | http -v :8080/klaus +echo -n Nachricht 2 an peter über producer | http -v :8080/peter +echo -n Nachricht 3 an peter über producer | http -v :8080/peter -docker compose -f docker/docker-compose.yml stop producer -docker compose -f docker/docker-compose.yml exec cli kafkacat -b kafka:9092 -t test -e -f'topic=%t\tpartition=%p\toffset=%o\tkey=%k\tvalue=%s\n' -docker compose -f docker/docker-compose.yml logs producer +echo Nachrichten in Partition 0: +kafkacat -b :9092 -t test -o 0 -p0 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +echo +echo Nachrichten in Partition 1: +kafkacat -b :9092 -t test -o 0 -p1 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +echo + +docker compose -f docker/docker-compose.yml exec -T cli bash << 'EOF' +echo "Altering number of partitions from 2 to 3..." +kafka-topics --bootstrap-server kafka:9092 --describe --topic test +# tag::repartitioning[] +kafka-topics --bootstrap-server kafka:9092 --alter --topic test --partitions 3 +# end::repartitioning[] +kafka-topics --bootstrap-server kafka:9092 --describe --topic test +EOF + +docker compose -f docker/docker-compose.yml restart producer +while ! [[ $(http 0:8080/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer..."; sleep 1; done + +echo -n Nachricht 4 an klaus über producer | http -v :8080/klaus +echo -n Nachricht 5 an peter über producer | http -v :8080/peter +echo -n Nachricht 4 an peter über producer | http -v :8080/peter +echo -n Nachricht 5 an klaus über producer | http -v :8080/klaus +echo -n Nachricht 6 an klaus über producer | http -v :8080/klaus +echo -n Nachricht 6 an peter über producer | http -v :8080/peter + +echo Nachrichten in Partition 0: +# tag::kafkacat[] +kafkacat -b :9092 -t test -o 0 -p0 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +# end::kafkacat[] +echo +echo Nachrichten in Partition 1: +# tag::kafkacat[] +kafkacat -b :9092 -t test -o 0 -p1 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +# end::kafkacat[] +echo +echo Nachrichten in Partition 2: +kafkacat -b :9092 -t test -o 0 -p2 -f'key: %k\toffset: %o\tvalue: %s\n' -qe + + +docker compose -f docker/docker-compose.yml restart setup +sleep 1 +docker compose -f docker/docker-compose.yml up -d producer-0 producer-1 +while ! [[ $(http 0:8000/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer-0..."; sleep 1; done +while ! [[ $(http 0:8001/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer-1..."; sleep 1; done + +# tag::fixed[] +echo -n Nachricht 1 über producer-0 | http -v :8000/klaus +echo -n Nachricht 1 über producer-1 | http -v :8001/klaus +echo -n Nachricht 2 über producer-0 | http -v :8000/peter +echo -n Nachricht 2 über producer-1 | http -v :8001/peter +# end::fixed[] + +docker compose -f docker/docker-compose.yml exec -T cli bash << 'EOF' +echo "Altering number of partitions from 2 to 3..." +kafka-topics --bootstrap-server kafka:9092 --describe --topic test +kafka-topics --bootstrap-server kafka:9092 --alter --topic test --partitions 3 +kafka-topics --bootstrap-server kafka:9092 --describe --topic test +EOF + +docker compose -f docker/docker-compose.yml restart producer-0 producer-1 +while ! [[ $(http 0:8000/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer-0..."; sleep 1; done +while ! [[ $(http 0:8001/actuator/health 2> /dev/null) =~ "UP" ]]; do echo "Waiting for producer-1..."; sleep 1; done + +echo -n Nachricht 3 über producer-0 | http -v :8000/klaus +echo -n Nachricht 3 über producer-1 | http -v :8001/klaus +echo -n Nachricht 4 über producer-0 | http -v :8000/peter +echo -n Nachricht 4 über producer-1 | http -v :8001/peter + +echo Nachrichten in Partition 0: +kafkacat -b :9092 -t test -o 0 -p0 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +echo +echo Nachrichten in Partition 1: +kafkacat -b :9092 -t test -o 0 -p1 -f'key: %k\toffset: %o\tvalue: %s\n' -qe +echo +echo Nachrichten in Partition 2: +kafkacat -b :9092 -t test -o 0 -p2 -f'key: %k\toffset: %o\tvalue: %s\n' -qe diff --git a/docker/docker-compose.yml b/docker/docker-compose.yml index d03918d..2eaa6b6 100644 --- a/docker/docker-compose.yml +++ b/docker/docker-compose.yml @@ -188,12 +188,49 @@ services: - kafka-3 producer: - image: juplo/simple-producer:1.0-SNAPSHOT + image: juplo/rest-producer:1.0-SNAPSHOT + ports: + - 8080:8080 environment: + server.port: 8080 producer.bootstrap-server: kafka:9092 producer.client-id: producer producer.topic: test + producer-0: + image: juplo/rest-producer:1.0-SNAPSHOT + ports: + - 8000:8080 + environment: + server.port: 8080 + producer.bootstrap-server: kafka:9092 + producer.client-id: producer-0 + producer.topic: test + producer.partition: 0 + + producer-1: + image: juplo/rest-producer:1.0-SNAPSHOT + ports: + - 8001:8080 + environment: + server.port: 8080 + producer.bootstrap-server: kafka:9092 + producer.client-id: producer-1 + producer.topic: test + producer.partition: 1 + + consumer-1: + image: juplo/simple-consumer:1.0-SNAPSHOT + command: kafka:9092 test my-group consumer-1 + + consumer-2: + image: juplo/simple-consumer:1.0-SNAPSHOT + command: kafka:9092 test my-group consumer-2 + + consumer-3: + image: juplo/simple-consumer:1.0-SNAPSHOT + command: kafka:9092 test my-group consumer-3 + volumes: zookeeper-data: zookeeper-log: diff --git a/pom.xml b/pom.xml index 841299b..999c66b 100644 --- a/pom.xml +++ b/pom.xml @@ -12,9 +12,9 @@ de.juplo.kafka - spring-producer - Spring Producer - A Simple Spring-Boot-Producer, that takes messages via POST and confirms successs + rest-producer + REST Producer + A Simple Producer that takes messages via POST and confirms successs 1.0-SNAPSHOT diff --git a/src/main/java/de/juplo/kafka/Application.java b/src/main/java/de/juplo/kafka/Application.java index 0069257..ba6aeee 100644 --- a/src/main/java/de/juplo/kafka/Application.java +++ b/src/main/java/de/juplo/kafka/Application.java @@ -2,9 +2,12 @@ package de.juplo.kafka; import org.springframework.boot.SpringApplication; import org.springframework.boot.autoconfigure.SpringBootApplication; +import org.springframework.context.annotation.ComponentScan; +import org.springframework.web.bind.annotation.RestController; @SpringBootApplication +@ComponentScan(excludeFilters = @ComponentScan.Filter(RestController.class)) public class Application { public static void main(String[] args) diff --git a/src/main/java/de/juplo/kafka/ApplicationConfiguration.java b/src/main/java/de/juplo/kafka/ApplicationConfiguration.java index 5211b12..9ffe585 100644 --- a/src/main/java/de/juplo/kafka/ApplicationConfiguration.java +++ b/src/main/java/de/juplo/kafka/ApplicationConfiguration.java @@ -22,6 +22,7 @@ public class ApplicationConfiguration new ExampleProducer( properties.getClientId(), properties.getTopic(), + properties.getPartition(), kafkaProducer); } diff --git a/src/main/java/de/juplo/kafka/ApplicationProperties.java b/src/main/java/de/juplo/kafka/ApplicationProperties.java index 4bf66a8..8efacd4 100644 --- a/src/main/java/de/juplo/kafka/ApplicationProperties.java +++ b/src/main/java/de/juplo/kafka/ApplicationProperties.java @@ -22,6 +22,7 @@ public class ApplicationProperties @NotNull @NotEmpty private String topic; + private Integer partition; @NotNull @NotEmpty private String acks; diff --git a/src/main/java/de/juplo/kafka/ErrorResponse.java b/src/main/java/de/juplo/kafka/ErrorResponse.java new file mode 100644 index 0000000..5ca206d --- /dev/null +++ b/src/main/java/de/juplo/kafka/ErrorResponse.java @@ -0,0 +1,11 @@ +package de.juplo.kafka; + +import lombok.Value; + + +@Value +public class ErrorResponse +{ + private final String error; + private final Integer status; +} diff --git a/src/main/java/de/juplo/kafka/ExampleProducer.java b/src/main/java/de/juplo/kafka/ExampleProducer.java index bbe014e..94608f9 100644 --- a/src/main/java/de/juplo/kafka/ExampleProducer.java +++ b/src/main/java/de/juplo/kafka/ExampleProducer.java @@ -1,71 +1,51 @@ package de.juplo.kafka; +import lombok.RequiredArgsConstructor; import lombok.extern.slf4j.Slf4j; import org.apache.kafka.clients.producer.Producer; import org.apache.kafka.clients.producer.ProducerRecord; +import org.springframework.http.HttpStatus; +import org.springframework.web.bind.annotation.*; +import org.springframework.web.context.request.async.DeferredResult; -import java.util.concurrent.Callable; +import java.math.BigInteger; @Slf4j -public class ExampleProducer implements Runnable +@RestController +@RequiredArgsConstructor +public class ExampleProducer { private final String id; private final String topic; + private final Integer partition; private final Producer producer; - private final Thread workerThread; - private volatile boolean running = true; private long produced = 0; - - public ExampleProducer( - String id, - String topic, - Producer producer) - { - this.id = id; - this.topic = topic; - this.producer = producer; - - workerThread = new Thread(this, "ExampleProducer Worker-Thread"); - workerThread.start(); - } - - - @Override - public void run() + @PostMapping(path = "{key}") + public DeferredResult send( + @PathVariable String key, + @RequestHeader(name = "X-id", required = false) Long correlationId, + @RequestBody String value) { - long i = 0; - - try - { - for (; running; i++) - { - send(Long.toString(i%10), Long.toString(i)); - Thread.sleep(500); - } - } - catch (Exception e) - { - log.error("{} - Unexpected error: {}!", id, e.toString()); - } - finally - { - log.info("{}: Produced {} messages in total, exiting!", id, produced); - } - } + DeferredResult result = new DeferredResult<>(); - void send(String key, String value) - { final long time = System.currentTimeMillis(); final ProducerRecord record = new ProducerRecord<>( topic, // Topic + partition, // Partition key, // Key value // Value ); + record.headers().add("source", id.getBytes()); + if (correlationId != null) + { + record.headers().add("id", BigInteger.valueOf(correlationId).toByteArray()); + } + producer.send(record, (metadata, e) -> { long now = System.currentTimeMillis(); @@ -73,6 +53,7 @@ public class ExampleProducer implements Runnable { // HANDLE SUCCESS produced++; + result.setResult(new ProduceSuccess(metadata.partition(), metadata.offset())); log.debug( "{} - Sent key={} message={} partition={}/{} timestamp={} latency={}ms", id, @@ -87,6 +68,7 @@ public class ExampleProducer implements Runnable else { // HANDLE ERROR + result.setErrorResult(new ProduceFailure(e)); log.error( "{} - ERROR key={} timestamp={} latency={}ms: {}", id, @@ -105,13 +87,14 @@ public class ExampleProducer implements Runnable record.key(), now - time ); - } + return result; + } - public void shutdown() throws InterruptedException + @ExceptionHandler + @ResponseStatus(HttpStatus.BAD_REQUEST) + public ErrorResponse illegalStateException(IllegalStateException e) { - log.info("{} joining the worker-thread...", id); - running = false; - workerThread.join(); + return new ErrorResponse(e.getMessage(), HttpStatus.BAD_REQUEST.value()); } } diff --git a/src/main/java/de/juplo/kafka/ProduceFailure.java b/src/main/java/de/juplo/kafka/ProduceFailure.java new file mode 100644 index 0000000..873a67b --- /dev/null +++ b/src/main/java/de/juplo/kafka/ProduceFailure.java @@ -0,0 +1,21 @@ +package de.juplo.kafka; + + +import lombok.Value; + + +@Value +public class ProduceFailure implements ProduceResult +{ + private final String error; + private final String exception; + private final Integer status; + + + public ProduceFailure(Exception e) + { + status = 500; + exception = e.getClass().getSimpleName(); + error = e.getMessage(); + } +} diff --git a/src/main/java/de/juplo/kafka/ProduceResult.java b/src/main/java/de/juplo/kafka/ProduceResult.java new file mode 100644 index 0000000..ceff329 --- /dev/null +++ b/src/main/java/de/juplo/kafka/ProduceResult.java @@ -0,0 +1,11 @@ +package de.juplo.kafka; + +import com.fasterxml.jackson.annotation.JsonInclude; + +import static com.fasterxml.jackson.annotation.JsonInclude.Include.NON_NULL; + + +@JsonInclude(NON_NULL) +public interface ProduceResult +{ +} diff --git a/src/main/java/de/juplo/kafka/ProduceSuccess.java b/src/main/java/de/juplo/kafka/ProduceSuccess.java new file mode 100644 index 0000000..9c79e8b --- /dev/null +++ b/src/main/java/de/juplo/kafka/ProduceSuccess.java @@ -0,0 +1,12 @@ +package de.juplo.kafka; + + +import lombok.Value; + + +@Value +public class ProduceSuccess implements ProduceResult +{ + Integer partition; + Long offset; +} diff --git a/src/test/java/de/juplo/kafka/ApplicationTests.java b/src/test/java/de/juplo/kafka/ApplicationTests.java index 8d579e9..50bc20b 100644 --- a/src/test/java/de/juplo/kafka/ApplicationTests.java +++ b/src/test/java/de/juplo/kafka/ApplicationTests.java @@ -27,6 +27,7 @@ import static org.springframework.test.web.servlet.result.MockMvcResultMatchers. properties = { "spring.kafka.consumer.bootstrap-servers=${spring.embedded.kafka.brokers}", "producer.bootstrap-server=${spring.embedded.kafka.brokers}", + "spring.kafka.consumer.auto-offset-reset=earliest", "producer.topic=" + TOPIC}) @AutoConfigureMockMvc @EmbeddedKafka(topics = TOPIC, partitions = PARTITIONS) @@ -36,6 +37,8 @@ public class ApplicationTests static final String TOPIC = "FOO"; static final int PARTITIONS = 10; + @Autowired + MockMvc mockMvc; @Autowired Consumer consumer; @@ -50,9 +53,12 @@ public class ApplicationTests @Test void testSendMessage() throws Exception { - await("Some messages were send") + mockMvc + .perform(post("/peter").content("Hallo Welt!")) + .andExpect(status().isOk()); + await("Message was send") .atMost(Duration.ofSeconds(5)) - .until(() -> consumer.received.size() >= 1); + .until(() -> consumer.received.size() == 1); } -- 2.20.1