package de.juplo.kafka;
import lombok.extern.slf4j.Slf4j;
+import org.apache.kafka.clients.consumer.ConsumerRebalanceListener;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
+import org.apache.kafka.common.TopicPartition;
import org.apache.kafka.common.errors.WakeupException;
import org.apache.kafka.common.serialization.StringDeserializer;
import javax.annotation.PreDestroy;
import java.time.Duration;
-import java.util.Arrays;
-import java.util.HashMap;
-import java.util.Map;
-import java.util.Optional;
-import java.util.Properties;
+import java.util.*;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.locks.Condition;
public class EndlessConsumer implements Runnable
{
private final ExecutorService executor;
+ private final PartitionStatisticsRepository repository;
private final String bootstrapServer;
private final String groupId;
private final String id;
private KafkaConsumer<String, String> consumer = null;
- private Map<Integer, Map<String, Integer>> seen;
+ private final Map<Integer, Map<String, Integer>> seen = new HashMap<>();
public EndlessConsumer(
ExecutorService executor,
+ PartitionStatisticsRepository repository,
String bootstrapServer,
String groupId,
String clientId,
String autoOffsetReset)
{
this.executor = executor;
+ this.repository = repository;
this.bootstrapServer = bootstrapServer;
this.groupId = groupId;
this.id = clientId;
this.consumer = new KafkaConsumer<>(props);
log.info("{} - Subscribing to topic {}", id, topic);
- consumer.subscribe(Arrays.asList(topic));
+ consumer.subscribe(Arrays.asList(topic), new ConsumerRebalanceListener()
+ {
+ @Override
+ public void onPartitionsRevoked(Collection<TopicPartition> partitions)
+ {
+ partitions.forEach(tp ->
+ {
+ log.info("{} - removing partition: {}", id, tp);
+ Map<String, Integer> removed = seen.remove(tp.partition());
+ for (String key : removed.keySet())
+ {
+ log.info(
+ "{} - Seen {} messages for partition={}|key={}",
+ id,
+ removed.get(key),
+ tp.partition(),
+ key);
+ }
+ repository.save(new StatisticsDocument(tp.partition(), removed));
+ });
+ }
- seen = new HashMap<>();
+ @Override
+ public void onPartitionsAssigned(Collection<TopicPartition> partitions)
+ {
+ partitions.forEach(tp ->
+ {
+ log.info("{} - adding partition: {}", id, tp);
+ seen.put(
+ tp.partition(),
+ repository
+ .findById(Integer.toString(tp.partition()))
+ .map(document -> document.statistics)
+ .orElse(new HashMap<>()));
+ });
+ }
+ });
while (true)
{
Integer partition = record.partition();
String key = record.key() == null ? "NULL" : record.key();
-
- if (!seen.containsKey(partition))
- seen.put(partition, new HashMap<>());
-
Map<String, Integer> byKey = seen.get(partition);
if (!byKey.containsKey(key))
{
log.info("{} - Closing the KafkaConsumer", id);
consumer.close();
-
- for (Integer partition : seen.keySet())
- {
- Map<String, Integer> byKey = seen.get(partition);
- for (String key : byKey.keySet())
- {
- log.info(
- "{} - Seen {} messages for partition={}|key={}",
- id,
- byKey.get(key),
- partition,
- key);
- }
- }
- seen = null;
-
log.info("{} - Consumer-Thread exiting", id);
}
}
- public Map<Integer, Map<String, Integer>> getSeen()
- {
- return seen;
- }
-
private void shutdown()
{
shutdown(null);
}
}
+ public Map<Integer, Map<String, Integer>> getSeen()
+ {
+ return seen;
+ }
+
public void start()
{
lock.lock();