X-Git-Url: http://juplo.de/gitweb/?a=blobdiff_plain;f=src%2Fmain%2Fjava%2Fde%2Fjuplo%2Fkafka%2FApplicationConfiguration.java;h=4923b09ec1bad97610416c26b659ebdf0432fa1f;hb=1709f0e4f41be7e3b955d19769697a517633827d;hp=3c526dfa0efb8a55c683028bae34bd98d7da71a5;hpb=d8f4e839b57b8405ca909b5472a8d4639937a101;p=demos%2Fkafka%2Ftraining diff --git a/src/main/java/de/juplo/kafka/ApplicationConfiguration.java b/src/main/java/de/juplo/kafka/ApplicationConfiguration.java index 3c526df..4923b09 100644 --- a/src/main/java/de/juplo/kafka/ApplicationConfiguration.java +++ b/src/main/java/de/juplo/kafka/ApplicationConfiguration.java @@ -1,17 +1,21 @@ package de.juplo.kafka; import org.apache.kafka.clients.consumer.ConsumerRecord; -import org.apache.kafka.clients.consumer.KafkaConsumer; -import org.apache.kafka.common.serialization.LongDeserializer; -import org.apache.kafka.common.serialization.StringDeserializer; +import org.apache.kafka.common.TopicPartition; +import org.apache.kafka.common.serialization.ByteArraySerializer; +import org.apache.kafka.common.serialization.StringSerializer; import org.springframework.boot.autoconfigure.kafka.KafkaProperties; import org.springframework.boot.context.properties.EnableConfigurationProperties; import org.springframework.context.annotation.Bean; import org.springframework.context.annotation.Configuration; +import org.springframework.kafka.core.*; +import org.springframework.kafka.listener.DeadLetterPublishingRecoverer; +import org.springframework.kafka.listener.DefaultErrorHandler; +import org.springframework.kafka.support.serializer.DelegatingByTypeSerializer; +import org.springframework.kafka.support.serializer.JsonSerializer; +import org.springframework.util.backoff.FixedBackOff; -import java.util.Properties; -import java.util.concurrent.ExecutorService; -import java.util.concurrent.Executors; +import java.util.Map; import java.util.function.Consumer; @@ -20,7 +24,7 @@ import java.util.function.Consumer; public class ApplicationConfiguration { @Bean - public Consumer> consumer() + public Consumer> consumer() { return (record) -> { @@ -29,41 +33,41 @@ public class ApplicationConfiguration } @Bean - public EndlessConsumer endlessConsumer( - KafkaConsumer kafkaConsumer, - ExecutorService executor, - Consumer> handler, - KafkaProperties kafkaProperties, - ApplicationProperties applicationProperties) + public ProducerFactory producerFactory(KafkaProperties properties) { + return new DefaultKafkaProducerFactory<>( + properties.getProducer().buildProperties(), + new StringSerializer(), + new DelegatingByTypeSerializer(Map.of( + byte[].class, new ByteArraySerializer(), + ClientMessage.class, new JsonSerializer<>()))); + } + + @Bean + public KafkaTemplate kafkaTemplate( + ProducerFactory producerFactory) { + + return new KafkaTemplate<>(producerFactory); + } + + @Bean + public DeadLetterPublishingRecoverer recoverer( + ApplicationProperties properties, + KafkaOperations template) { - return - new EndlessConsumer<>( - executor, - kafkaProperties.getConsumer().getClientId(), - applicationProperties.getTopic(), - kafkaConsumer, - handler); + return new DeadLetterPublishingRecoverer( + template, + (record, exception) -> new TopicPartition(properties.getDlqTopic(), record.partition())); } @Bean - public ExecutorService executor() + public DefaultErrorHandler errorHandler(DeadLetterPublishingRecoverer recoverer) { - return Executors.newSingleThreadExecutor(); + return new DefaultErrorHandler(recoverer, new FixedBackOff(0l, 0l)); } @Bean(destroyMethod = "close") - public KafkaConsumer kafkaConsumer(KafkaProperties properties) + public org.apache.kafka.clients.consumer.Consumer kafkaConsumer(ConsumerFactory factory) { - Properties props = new Properties(); - - props.put("bootstrap.servers", properties.getConsumer().getBootstrapServers()); - props.put("group.id", properties.getConsumer().getGroupId()); - props.put("client.id", properties.getConsumer().getClientId()); - props.put("auto.offset.reset", properties.getConsumer().getAutoOffsetReset()); - props.put("metadata.max.age.ms", "1000"); - props.put("key.deserializer", StringDeserializer.class.getName()); - props.put("value.deserializer", LongDeserializer.class.getName()); - - return new KafkaConsumer<>(props); + return factory.createConsumer(); } }