Die Key-Statistiken werden in einer MongoDB gespeichert
[demos/kafka/training] / src / main / java / de / juplo / kafka / EndlessConsumer.java
index da2f8f0..e67bf41 100644 (file)
@@ -1,16 +1,17 @@
 package de.juplo.kafka;
 
 import lombok.extern.slf4j.Slf4j;
+import org.apache.kafka.clients.consumer.ConsumerRebalanceListener;
 import org.apache.kafka.clients.consumer.ConsumerRecord;
 import org.apache.kafka.clients.consumer.ConsumerRecords;
 import org.apache.kafka.clients.consumer.KafkaConsumer;
+import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.errors.WakeupException;
 import org.apache.kafka.common.serialization.StringDeserializer;
 
 import javax.annotation.PreDestroy;
 import java.time.Duration;
-import java.util.Arrays;
-import java.util.Properties;
+import java.util.*;
 import java.util.concurrent.ExecutionException;
 import java.util.concurrent.ExecutorService;
 import java.util.concurrent.Future;
@@ -21,28 +22,37 @@ import java.util.concurrent.atomic.AtomicBoolean;
 public class EndlessConsumer implements Runnable
 {
   private final ExecutorService executor;
+  private final PartitionStatisticsRepository repository;
   private final String bootstrapServer;
   private final String groupId;
   private final String id;
   private final String topic;
+  private final String autoOffsetReset;
 
   private AtomicBoolean running = new AtomicBoolean();
   private long consumed = 0;
   private KafkaConsumer<String, String> consumer = null;
   private Future<?> future = null;
 
+  private final Map<TopicPartition, PartitionStatistics> seen = new HashMap<>();
+
+
   public EndlessConsumer(
       ExecutorService executor,
+      PartitionStatisticsRepository repository,
       String bootstrapServer,
       String groupId,
       String clientId,
-      String topic)
+      String topic,
+      String autoOffsetReset)
   {
     this.executor = executor;
+    this.repository = repository;
     this.bootstrapServer = bootstrapServer;
     this.groupId = groupId;
     this.id = clientId;
     this.topic = topic;
+    this.autoOffsetReset = autoOffsetReset;
   }
 
   @Override
@@ -54,14 +64,51 @@ public class EndlessConsumer implements Runnable
       props.put("bootstrap.servers", bootstrapServer);
       props.put("group.id", groupId);
       props.put("client.id", id);
-      props.put("auto.offset.reset", "earliest");
+      props.put("auto.offset.reset", autoOffsetReset);
+      props.put("metadata.max.age.ms", "1000");
       props.put("key.deserializer", StringDeserializer.class.getName());
       props.put("value.deserializer", StringDeserializer.class.getName());
 
       this.consumer = new KafkaConsumer<>(props);
 
       log.info("{} - Subscribing to topic {}", id, topic);
-      consumer.subscribe(Arrays.asList(topic));
+      consumer.subscribe(Arrays.asList(topic), new ConsumerRebalanceListener()
+      {
+        @Override
+        public void onPartitionsRevoked(Collection<TopicPartition> partitions)
+        {
+          partitions.forEach(tp ->
+          {
+            log.info("{} - removing partition: {}", id, tp);
+            PartitionStatistics removed = seen.remove(tp);
+            for (KeyCounter counter : removed.getStatistics())
+            {
+              log.info(
+                  "{} - Seen {} messages for partition={}|key={}",
+                  id,
+                  counter.getResult(),
+                  removed.getPartition(),
+                  counter.getKey());
+            }
+            repository.save(new StatisticsDocument(removed));
+          });
+        }
+
+        @Override
+        public void onPartitionsAssigned(Collection<TopicPartition> partitions)
+        {
+          partitions.forEach(tp ->
+          {
+            log.info("{} - adding partition: {}", id, tp);
+            seen.put(
+                tp,
+                repository
+                    .findById(tp.toString())
+                    .map(PartitionStatistics::new)
+                    .orElse(new PartitionStatistics(tp)));
+          });
+        }
+      });
 
       while (true)
       {
@@ -82,6 +129,10 @@ public class EndlessConsumer implements Runnable
               record.key(),
               record.value()
           );
+
+          TopicPartition partition = new TopicPartition(record.topic(), record.partition());
+          String key = record.key() == null ? "NULL" : record.key();
+          seen.get(partition).increment(key);
         }
       }
     }
@@ -102,6 +153,10 @@ public class EndlessConsumer implements Runnable
     }
   }
 
+  public Map<TopicPartition, PartitionStatistics> getSeen()
+  {
+    return seen;
+  }
 
   public synchronized void start()
   {