X-Git-Url: https://juplo.de/gitweb/?a=blobdiff_plain;f=src%2Fmain%2Fjava%2Fde%2Fjuplo%2Fkafka%2FEndlessConsumer.java;h=e67bf412a1bae6762c7270935079e47f57716a1f;hb=915674ec49ba38b3716cc4ef53272e963f139677;hp=b3dd446d84ddbc7fddd44d516cbd921eff1ece7e;hpb=6ce3e7f2a3ff4622a82e4bb1ec35a722d02e2069;p=demos%2Fkafka%2Ftraining diff --git a/src/main/java/de/juplo/kafka/EndlessConsumer.java b/src/main/java/de/juplo/kafka/EndlessConsumer.java index b3dd446..e67bf41 100644 --- a/src/main/java/de/juplo/kafka/EndlessConsumer.java +++ b/src/main/java/de/juplo/kafka/EndlessConsumer.java @@ -1,16 +1,17 @@ package de.juplo.kafka; import lombok.extern.slf4j.Slf4j; +import org.apache.kafka.clients.consumer.ConsumerRebalanceListener; import org.apache.kafka.clients.consumer.ConsumerRecord; import org.apache.kafka.clients.consumer.ConsumerRecords; import org.apache.kafka.clients.consumer.KafkaConsumer; +import org.apache.kafka.common.TopicPartition; import org.apache.kafka.common.errors.WakeupException; import org.apache.kafka.common.serialization.StringDeserializer; import javax.annotation.PreDestroy; import java.time.Duration; -import java.util.Arrays; -import java.util.Properties; +import java.util.*; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorService; import java.util.concurrent.Future; @@ -21,6 +22,7 @@ import java.util.concurrent.atomic.AtomicBoolean; public class EndlessConsumer implements Runnable { private final ExecutorService executor; + private final PartitionStatisticsRepository repository; private final String bootstrapServer; private final String groupId; private final String id; @@ -32,8 +34,12 @@ public class EndlessConsumer implements Runnable private KafkaConsumer consumer = null; private Future future = null; + private final Map seen = new HashMap<>(); + + public EndlessConsumer( ExecutorService executor, + PartitionStatisticsRepository repository, String bootstrapServer, String groupId, String clientId, @@ -41,6 +47,7 @@ public class EndlessConsumer implements Runnable String autoOffsetReset) { this.executor = executor; + this.repository = repository; this.bootstrapServer = bootstrapServer; this.groupId = groupId; this.id = clientId; @@ -58,13 +65,50 @@ public class EndlessConsumer implements Runnable props.put("group.id", groupId); props.put("client.id", id); props.put("auto.offset.reset", autoOffsetReset); + props.put("metadata.max.age.ms", "1000"); props.put("key.deserializer", StringDeserializer.class.getName()); props.put("value.deserializer", StringDeserializer.class.getName()); this.consumer = new KafkaConsumer<>(props); log.info("{} - Subscribing to topic {}", id, topic); - consumer.subscribe(Arrays.asList(topic)); + consumer.subscribe(Arrays.asList(topic), new ConsumerRebalanceListener() + { + @Override + public void onPartitionsRevoked(Collection partitions) + { + partitions.forEach(tp -> + { + log.info("{} - removing partition: {}", id, tp); + PartitionStatistics removed = seen.remove(tp); + for (KeyCounter counter : removed.getStatistics()) + { + log.info( + "{} - Seen {} messages for partition={}|key={}", + id, + counter.getResult(), + removed.getPartition(), + counter.getKey()); + } + repository.save(new StatisticsDocument(removed)); + }); + } + + @Override + public void onPartitionsAssigned(Collection partitions) + { + partitions.forEach(tp -> + { + log.info("{} - adding partition: {}", id, tp); + seen.put( + tp, + repository + .findById(tp.toString()) + .map(PartitionStatistics::new) + .orElse(new PartitionStatistics(tp))); + }); + } + }); while (true) { @@ -85,6 +129,10 @@ public class EndlessConsumer implements Runnable record.key(), record.value() ); + + TopicPartition partition = new TopicPartition(record.topic(), record.partition()); + String key = record.key() == null ? "NULL" : record.key(); + seen.get(partition).increment(key); } } } @@ -105,6 +153,10 @@ public class EndlessConsumer implements Runnable } } + public Map getSeen() + { + return seen; + } public synchronized void start() {