X-Git-Url: https://juplo.de/gitweb/?a=blobdiff_plain;f=src%2Fmain%2Fjava%2Fde%2Fjuplo%2Fkafka%2FWordcountRebalanceListener.java;h=9f2fc0f75dc193dfb72438f01a970d857e63875b;hb=0b6377951bbeeb2443415bc2aa9c68988ee4c5af;hp=fd551c2302d9a969026f5686459bbc6f48dbf776;hpb=818c1eb862247e25abf9f7d91d5a73e3e3789a39;p=demos%2Fkafka%2Ftraining diff --git a/src/main/java/de/juplo/kafka/WordcountRebalanceListener.java b/src/main/java/de/juplo/kafka/WordcountRebalanceListener.java index fd551c2..9f2fc0f 100644 --- a/src/main/java/de/juplo/kafka/WordcountRebalanceListener.java +++ b/src/main/java/de/juplo/kafka/WordcountRebalanceListener.java @@ -1,27 +1,83 @@ package de.juplo.kafka; import lombok.RequiredArgsConstructor; -import org.apache.kafka.clients.consumer.ConsumerRebalanceListener; +import lombok.extern.slf4j.Slf4j; +import org.apache.kafka.clients.consumer.Consumer; import org.apache.kafka.common.TopicPartition; +import java.time.Clock; +import java.time.Duration; +import java.time.Instant; import java.util.Collection; +import java.util.Map; @RequiredArgsConstructor -public class WordcountRebalanceListener implements ConsumerRebalanceListener +@Slf4j +public class WordcountRebalanceListener implements PollIntervalAwareConsumerRebalanceListener { - private final RecordHandler handler; + private final WordcountRecordHandler handler; + private final PartitionStatisticsRepository repository; + private final String id; + private final String topic; + private final Clock clock; + private final Duration commitInterval; + private final Consumer consumer; + private Instant lastCommit = Instant.EPOCH; @Override public void onPartitionsAssigned(Collection partitions) { - partitions.forEach(tp -> handler.onPartitionAssigned(tp)); + partitions.forEach(tp -> + { + Integer partition = tp.partition(); + Long offset = consumer.position(tp); + log.info("{} - adding partition: {}, offset={}", id, partition, offset); + StatisticsDocument document = + repository + .findById(Integer.toString(partition)) + .orElse(new StatisticsDocument(partition)); + if (document.offset >= 0) + { + // Only seek, if a stored offset was found + // Otherwise: Use initial offset, generated by Kafka + consumer.seek(tp, document.offset); + } + handler.addPartition(partition, document.statistics); + }); } @Override public void onPartitionsRevoked(Collection partitions) { - partitions.forEach(tp -> handler.onPartitionRevoked(tp)); + partitions.forEach(tp -> + { + Integer partition = tp.partition(); + Long newOffset = consumer.position(tp); + log.info( + "{} - removing partition: {}, offset of next message {})", + id, + partition, + newOffset); + Map> removed = handler.removePartition(partition); + repository.save(new StatisticsDocument(partition, removed, consumer.position(tp))); + }); + } + + + @Override + public void beforeNextPoll() + { + if (lastCommit.plus(commitInterval).isBefore(clock.instant())) + { + log.debug("Storing data and offsets, last commit: {}", lastCommit); + handler.getSeen().forEach((partiton, statistics) -> repository.save( + new StatisticsDocument( + partiton, + statistics, + consumer.position(new TopicPartition(topic, partiton))))); + lastCommit = clock.instant(); + } } }