projects
/
demos
/
kafka
/
training
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Der Test verwendet die `@Bean` von `EndlessConsumer`
[demos/kafka/training]
/
src
/
test
/
java
/
de
/
juplo
/
kafka
/
TestRecordHandler.java
diff --git
a/src/test/java/de/juplo/kafka/TestRecordHandler.java
b/src/test/java/de/juplo/kafka/TestRecordHandler.java
index
b4efdd6
..
37d3f65
100644
(file)
--- a/
src/test/java/de/juplo/kafka/TestRecordHandler.java
+++ b/
src/test/java/de/juplo/kafka/TestRecordHandler.java
@@
-2,16
+2,28
@@
package de.juplo.kafka;
import lombok.RequiredArgsConstructor;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import lombok.RequiredArgsConstructor;
import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.common.TopicPartition;
+
+import java.util.Map;
+import java.util.Set;
@RequiredArgsConstructor
@RequiredArgsConstructor
-public
abstract
class TestRecordHandler<K, V> implements RecordHandler<K, V>
+public class TestRecordHandler<K, V> implements RecordHandler<K, V>
{
private final RecordHandler<K, V> handler;
{
private final RecordHandler<K, V> handler;
+ Map<TopicPartition, Long> seenOffsets;
+ Set<ConsumerRecord<K, V>> receivedRecords;
- public abstract void onNewRecord(ConsumerRecord<K, V> record);
+ public void onNewRecord(ConsumerRecord<K, V> record)
+ {
+ seenOffsets.put(
+ new TopicPartition(record.topic(), record.partition()),
+ record.offset());
+ receivedRecords.add(record);
+ }
@Override
public void accept(ConsumerRecord<K, V> record)
@Override
public void accept(ConsumerRecord<K, V> record)