projects
/
demos
/
kafka
/
training
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge branch 'stored-state' into stored-offsets
[demos/kafka/training]
/
src
/
main
/
java
/
de
/
juplo
/
kafka
/
EndlessConsumer.java
diff --git
a/src/main/java/de/juplo/kafka/EndlessConsumer.java
b/src/main/java/de/juplo/kafka/EndlessConsumer.java
index
14a875b
..
2563204
100644
(file)
--- a/
src/main/java/de/juplo/kafka/EndlessConsumer.java
+++ b/
src/main/java/de/juplo/kafka/EndlessConsumer.java
@@
-22,6
+22,7
@@
import java.util.concurrent.atomic.AtomicBoolean;
public class EndlessConsumer implements Runnable
{
private final ExecutorService executor;
public class EndlessConsumer implements Runnable
{
private final ExecutorService executor;
+ private final PartitionStatisticsRepository repository;
private final String bootstrapServer;
private final String groupId;
private final String id;
private final String bootstrapServer;
private final String groupId;
private final String id;
@@
-38,6
+39,7
@@
public class EndlessConsumer implements Runnable
public EndlessConsumer(
ExecutorService executor,
public EndlessConsumer(
ExecutorService executor,
+ PartitionStatisticsRepository repository,
String bootstrapServer,
String groupId,
String clientId,
String bootstrapServer,
String groupId,
String clientId,
@@
-45,6
+47,7
@@
public class EndlessConsumer implements Runnable
String autoOffsetReset)
{
this.executor = executor;
String autoOffsetReset)
{
this.executor = executor;
+ this.repository = repository;
this.bootstrapServer = bootstrapServer;
this.groupId = groupId;
this.id = clientId;
this.bootstrapServer = bootstrapServer;
this.groupId = groupId;
this.id = clientId;
@@
-61,6
+64,7
@@
public class EndlessConsumer implements Runnable
props.put("bootstrap.servers", bootstrapServer);
props.put("group.id", groupId);
props.put("client.id", id);
props.put("bootstrap.servers", bootstrapServer);
props.put("group.id", groupId);
props.put("client.id", id);
+ props.put("enable.auto.commit", false);
props.put("auto.offset.reset", autoOffsetReset);
props.put("metadata.max.age.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
props.put("auto.offset.reset", autoOffsetReset);
props.put("metadata.max.age.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
@@
-87,6
+91,7
@@
public class EndlessConsumer implements Runnable
tp.partition(),
key);
}
tp.partition(),
key);
}
+ repository.save(new StatisticsDocument(tp.partition(), removed, consumer.position(tp)));
});
}
});
}
@@
-96,7
+101,12
@@
public class EndlessConsumer implements Runnable
partitions.forEach(tp ->
{
log.info("{} - adding partition: {}", id, tp);
partitions.forEach(tp ->
{
log.info("{} - adding partition: {}", id, tp);
- seen.put(tp.partition(), new HashMap<>());
+ StatisticsDocument document =
+ repository
+ .findById(Integer.toString(tp.partition()))
+ .orElse(new StatisticsDocument(tp.partition()));
+ consumer.seek(tp, document.offset);
+ seen.put(tp.partition(), document.statistics);
});
}
});
});
}
});
@@
-132,6
+142,12
@@
public class EndlessConsumer implements Runnable
seenByKey++;
byKey.put(key, seenByKey);
}
seenByKey++;
byKey.put(key, seenByKey);
}
+
+ seen.forEach((partiton, statistics) -> repository.save(
+ new StatisticsDocument(
+ partiton,
+ statistics,
+ consumer.position(new TopicPartition(topic, partiton)))));
}
}
catch(WakeupException e)
}
}
catch(WakeupException e)