ROT: Korrigierten/Verbesserten Test und Überarbeitetes Setup gemerged
[demos/kafka/training] / src / test / java / de / juplo / kafka / GenericApplicationTests.java
index 449c389..b019373 100644 (file)
@@ -41,7 +41,7 @@ import static org.awaitility.Awaitility.*;
                properties = {
                                "sumup.adder.bootstrap-server=${spring.embedded.kafka.brokers}",
                                "sumup.adder.topic=" + TOPIC,
-                               "sumup.adder.commit-interval=1s",
+                               "sumup.adder.commit-interval=500ms",
                                "spring.mongodb.embedded.version=4.4.13" })
 @EmbeddedKafka(topics = TOPIC, partitions = PARTITIONS)
 @EnableAutoConfiguration
@@ -62,6 +62,8 @@ abstract class GenericApplicationTests<K, V>
        @Autowired
        ExecutorService executor;
        @Autowired
+       StateRepository stateRepository;
+       @Autowired
        MongoClient mongoClient;
        @Autowired
        MongoProperties mongoProperties;
@@ -177,9 +179,6 @@ abstract class GenericApplicationTests<K, V>
 
                checkSeenOffsetsForProgress();
                compareToCommitedOffsets(oldOffsets);
-               assertThat(receivedRecords.size())
-                               .describedAs("Received not all sent events")
-                               .isLessThan(numberOfGeneratedMessages);
 
                assertThatNoException()
                                .describedAs("Consumer should not be running")
@@ -231,23 +230,29 @@ abstract class GenericApplicationTests<K, V>
        void seekToEnd()
        {
                offsetConsumer.assign(partitions());
-               offsetConsumer.seekToEnd(partitions());
                partitions().forEach(tp ->
                {
-                       // seekToEnd() works lazily: it only takes effect on poll()/position()
                        Long offset = offsetConsumer.position(tp);
                        log.info("New position for {}: {}", tp, offset);
+                       Integer partition = tp.partition();
+                       StateDocument document =
+                                       stateRepository
+                                                       .findById(partition.toString())
+                                                       .orElse(new StateDocument(partition));
+                       document.offset = offset;
+                       stateRepository.save(document);
                });
-               // The new positions must be commited!
-               offsetConsumer.commitSync();
                offsetConsumer.unsubscribe();
        }
 
        void doForCurrentOffsets(BiConsumer<TopicPartition, Long> consumer)
        {
-               offsetConsumer.assign(partitions());
-               partitions().forEach(tp -> consumer.accept(tp, offsetConsumer.position(tp)));
-               offsetConsumer.unsubscribe();
+               partitions().forEach(tp ->
+               {
+                       String partition = Integer.toString(tp.partition());
+                       Optional<Long> offset = stateRepository.findById(partition).map(document -> document.offset);
+                       consumer.accept(tp, offset.orElse(0l));
+               });
        }
 
        List<TopicPartition> partitions()
@@ -327,6 +332,7 @@ abstract class GenericApplicationTests<K, V>
                props.put("value.deserializer", BytesDeserializer.class.getName());
                offsetConsumer = new KafkaConsumer<>(props);
 
+               mongoClient.getDatabase(mongoProperties.getDatabase()).drop();
                seekToEnd();
 
                oldOffsets = new HashMap<>();
@@ -352,8 +358,6 @@ abstract class GenericApplicationTests<K, V>
                                        }
                                };
 
-               mongoClient.getDatabase(mongoProperties.getDatabase()).drop();
-
                endlessConsumer =
                                new EndlessConsumer<>(
                                                executor,