GRÜN: (ungewollt!) - Unabhängigkeit der Tests wieder hergestellt
[demos/kafka/training] / src / test / java / de / juplo / kafka / GenericApplicationTests.java
index 8b9a3ff..449c389 100644 (file)
@@ -1,5 +1,6 @@
 package de.juplo.kafka;
 
+import com.mongodb.client.MongoClient;
 import lombok.extern.slf4j.Slf4j;
 import org.apache.kafka.clients.consumer.ConsumerRecord;
 import org.apache.kafka.clients.consumer.KafkaConsumer;
@@ -11,6 +12,9 @@ import org.apache.kafka.common.serialization.*;
 import org.apache.kafka.common.utils.Bytes;
 import org.junit.jupiter.api.*;
 import org.springframework.beans.factory.annotation.Autowired;
+import org.springframework.boot.autoconfigure.EnableAutoConfiguration;
+import org.springframework.boot.autoconfigure.mongo.MongoProperties;
+import org.springframework.boot.test.autoconfigure.data.mongo.AutoConfigureDataMongo;
 import org.springframework.boot.test.context.ConfigDataApplicationContextInitializer;
 import org.springframework.boot.test.context.TestConfiguration;
 import org.springframework.context.annotation.Import;
@@ -35,10 +39,13 @@ import static org.awaitility.Awaitility.*;
 @SpringJUnitConfig(initializers = ConfigDataApplicationContextInitializer.class)
 @TestPropertySource(
                properties = {
-                               "consumer.bootstrap-server=${spring.embedded.kafka.brokers}",
-                               "consumer.topic=" + TOPIC,
-                               "consumer.commit-interval=1s" })
+                               "sumup.adder.bootstrap-server=${spring.embedded.kafka.brokers}",
+                               "sumup.adder.topic=" + TOPIC,
+                               "sumup.adder.commit-interval=1s",
+                               "spring.mongodb.embedded.version=4.4.13" })
 @EmbeddedKafka(topics = TOPIC, partitions = PARTITIONS)
+@EnableAutoConfiguration
+@AutoConfigureDataMongo
 @Slf4j
 abstract class GenericApplicationTests<K, V>
 {
@@ -54,6 +61,14 @@ abstract class GenericApplicationTests<K, V>
        ApplicationProperties properties;
        @Autowired
        ExecutorService executor;
+       @Autowired
+       MongoClient mongoClient;
+       @Autowired
+       MongoProperties mongoProperties;
+       @Autowired
+       PollIntervalAwareConsumerRebalanceListener rebalanceListener;
+       @Autowired
+       RecordHandler<K, V> recordHandler;
 
        KafkaProducer<Bytes, Bytes> testRecordProducer;
        KafkaConsumer<Bytes, Bytes> offsetConsumer;
@@ -78,12 +93,13 @@ abstract class GenericApplicationTests<K, V>
        @Test
        void commitsCurrentOffsetsOnSuccess()
        {
-               recordGenerator.generate(100, Set.of(), Set.of(), messageSender);
+               int numberOfGeneratedMessages =
+                               recordGenerator.generate(false, false, messageSender);
 
-               await("100 records received")
+               await(numberOfGeneratedMessages + " records received")
                                .atMost(Duration.ofSeconds(30))
                                .pollInterval(Duration.ofSeconds(1))
-                               .until(() -> receivedRecords.size() >= 100);
+                               .until(() -> receivedRecords.size() >= numberOfGeneratedMessages);
 
                await("Offsets committed")
                                .atMost(Duration.ofSeconds(10))
@@ -97,13 +113,16 @@ abstract class GenericApplicationTests<K, V>
                assertThatExceptionOfType(IllegalStateException.class)
                                .isThrownBy(() -> endlessConsumer.exitStatus())
                                .describedAs("Consumer should still be running");
+
+               recordGenerator.assertBusinessLogic();
        }
 
        @Test
        @SkipWhenErrorCannotBeGenerated(poisonPill = true)
        void commitsOffsetOfErrorForReprocessingOnDeserializationError()
        {
-               recordGenerator.generate(100, Set.of(77), Set.of(), messageSender);
+               int numberOfGeneratedMessages =
+                               recordGenerator.generate(true, false, messageSender);
 
                await("Consumer failed")
                                .atMost(Duration.ofSeconds(30))
@@ -123,7 +142,7 @@ abstract class GenericApplicationTests<K, V>
                compareToCommitedOffsets(newOffsets);
                assertThat(receivedRecords.size())
                                .describedAs("Received not all sent events")
-                               .isLessThan(100);
+                               .isLessThan(numberOfGeneratedMessages);
 
                assertThatNoException()
                                .describedAs("Consumer should not be running")
@@ -131,13 +150,16 @@ abstract class GenericApplicationTests<K, V>
                assertThat(endlessConsumer.exitStatus())
                                .describedAs("Consumer should have exited abnormally")
                                .containsInstanceOf(RecordDeserializationException.class);
+
+               recordGenerator.assertBusinessLogic();
        }
 
        @Test
        @SkipWhenErrorCannotBeGenerated(logicError = true)
        void doesNotCommitOffsetsOnLogicError()
        {
-               recordGenerator.generate(100, Set.of(), Set.of(77), messageSender);
+               int numberOfGeneratedMessages =
+                               recordGenerator.generate(false, true, messageSender);
 
                await("Consumer failed")
                                .atMost(Duration.ofSeconds(30))
@@ -157,7 +179,7 @@ abstract class GenericApplicationTests<K, V>
                compareToCommitedOffsets(oldOffsets);
                assertThat(receivedRecords.size())
                                .describedAs("Received not all sent events")
-                               .isLessThan(100);
+                               .isLessThan(numberOfGeneratedMessages);
 
                assertThatNoException()
                                .describedAs("Consumer should not be running")
@@ -165,6 +187,8 @@ abstract class GenericApplicationTests<K, V>
                assertThat(endlessConsumer.exitStatus())
                                .describedAs("Consumer should have exited abnormally")
                                .containsInstanceOf(RuntimeException.class);
+
+               recordGenerator.assertBusinessLogic();
        }
 
 
@@ -238,13 +262,12 @@ abstract class GenericApplicationTests<K, V>
 
        public interface RecordGenerator
        {
-               void generate(
-                               int numberOfMessagesToGenerate,
-                               Set<Integer> poisonPills,
-                               Set<Integer> logicErrors,
+               int generate(
+                               boolean poisonPills,
+                               boolean logicErrors,
                                Consumer<ProducerRecord<Bytes, Bytes>> messageSender);
 
-               default boolean canGeneratePoisionPill()
+               default boolean canGeneratePoisonPill()
                {
                        return true;
                }
@@ -253,6 +276,11 @@ abstract class GenericApplicationTests<K, V>
                {
                        return true;
                }
+
+               default void assertBusinessLogic()
+               {
+                       log.debug("No business-logic to assert");
+               }
        }
 
        void sendMessage(ProducerRecord<Bytes, Bytes> record)
@@ -311,22 +339,28 @@ abstract class GenericApplicationTests<K, V>
                        newOffsets.put(tp, offset - 1);
                });
 
-               Consumer<ConsumerRecord<K, V>> captureOffsetAndExecuteTestHandler =
-                               record ->
+               TestRecordHandler<K, V> captureOffsetAndExecuteTestHandler =
+                               new TestRecordHandler<K, V>(recordHandler)
                                {
-                                       newOffsets.put(
-                                                       new TopicPartition(record.topic(), record.partition()),
-                                                       record.offset());
-                                       receivedRecords.add(record);
-                                       consumer.accept(record);
+                                       @Override
+                                       public void onNewRecord(ConsumerRecord<K, V> record)
+                                       {
+                                               newOffsets.put(
+                                                               new TopicPartition(record.topic(), record.partition()),
+                                                               record.offset());
+                                               receivedRecords.add(record);
+                                       }
                                };
 
+               mongoClient.getDatabase(mongoProperties.getDatabase()).drop();
+
                endlessConsumer =
                                new EndlessConsumer<>(
                                                executor,
                                                properties.getClientId(),
                                                properties.getTopic(),
                                                kafkaConsumer,
+                                               rebalanceListener,
                                                captureOffsetAndExecuteTestHandler);
 
                endlessConsumer.start();