Springify: Merge des verschärften Tests aus der Vanilla-Version
[demos/kafka/training] / src / test / java / de / juplo / kafka / ApplicationTests.java
index 3ded0d2..1d3546c 100644 (file)
@@ -6,11 +6,10 @@ import org.apache.kafka.clients.consumer.KafkaConsumer;
 import org.apache.kafka.clients.producer.KafkaProducer;
 import org.apache.kafka.clients.producer.ProducerRecord;
 import org.apache.kafka.common.TopicPartition;
-import org.apache.kafka.common.serialization.BytesDeserializer;
-import org.apache.kafka.common.serialization.BytesSerializer;
-import org.apache.kafka.common.serialization.LongSerializer;
-import org.apache.kafka.common.serialization.StringSerializer;
+import org.apache.kafka.common.errors.RecordDeserializationException;
+import org.apache.kafka.common.serialization.*;
 import org.apache.kafka.common.utils.Bytes;
+import org.assertj.core.api.OptionalAssert;
 import org.junit.jupiter.api.*;
 import org.springframework.beans.factory.annotation.Autowired;
 import org.springframework.boot.autoconfigure.kafka.KafkaAutoConfiguration;
@@ -34,7 +33,7 @@ import java.util.stream.IntStream;
 
 import static de.juplo.kafka.ApplicationTests.PARTITIONS;
 import static de.juplo.kafka.ApplicationTests.TOPIC;
-import static org.assertj.core.api.Assertions.assertThat;
+import static org.assertj.core.api.Assertions.*;
 import static org.awaitility.Awaitility.*;
 
 
@@ -58,8 +57,9 @@ class ApplicationTests
 
 
        StringSerializer stringSerializer = new StringSerializer();
-       LongSerializer longSerializer = new LongSerializer();
 
+       @Autowired
+       Serializer valueSerializer;
        @Autowired
        KafkaProducer<String, Bytes> kafkaProducer;
        @Autowired
@@ -82,7 +82,7 @@ class ApplicationTests
        @Order(1) // << The poistion pill is not skipped. Hence, this test must run first
        void commitsCurrentOffsetsOnSuccess() throws ExecutionException, InterruptedException
        {
-               send100Messages(i ->  new Bytes(longSerializer.serialize(TOPIC, i)));
+               send100Messages(i ->  new Bytes(valueSerializer.serialize(TOPIC, i)));
 
                await("100 records received")
                                .atMost(Duration.ofSeconds(30))
@@ -95,6 +95,10 @@ class ApplicationTests
                                        checkSeenOffsetsForProgress();
                                        compareToCommitedOffsets(newOffsets);
                                });
+
+               assertThatExceptionOfType(IllegalStateException.class)
+                               .isThrownBy(() -> endlessConsumer.exitStatus())
+                               .describedAs("Consumer should still be running");
        }
 
        @Test
@@ -104,7 +108,7 @@ class ApplicationTests
                send100Messages(counter ->
                                counter == 77
                                                ? new Bytes(stringSerializer.serialize(TOPIC, "BOOM!"))
-                                               : new Bytes(longSerializer.serialize(TOPIC, counter)));
+                                               : new Bytes(valueSerializer.serialize(TOPIC, counter)));
 
                await("Consumer failed")
                                .atMost(Duration.ofSeconds(30))
@@ -114,6 +118,13 @@ class ApplicationTests
                assertThat(receivedRecords.size())
                                .describedAs("Received not all sent events")
                                .isLessThan(100);
+
+               assertThatNoException()
+                               .describedAs("Consumer should not be running")
+                               .isThrownBy(() -> endlessConsumer.exitStatus());
+               ((OptionalAssert<Exception>)assertThat(endlessConsumer.exitStatus()))
+                               .describedAs("Consumer should have exited abnormally")
+                               .containsInstanceOf(RecordDeserializationException.class);
        }
 
 
@@ -278,6 +289,12 @@ class ApplicationTests
                        return new RecordHandler();
                }
 
+               @Bean
+               Serializer<Long> serializer()
+               {
+                       return new LongSerializer();
+               }
+
                @Bean
                KafkaProducer<String, Bytes> kafkaProducer(ApplicationProperties properties)
                {