package de.juplo.kafka.wordcount.splitter;
-import de.juplo.kafka.wordcount.counter.TestWord;
-import de.juplo.kafka.wordcount.recorder.TestRecording;
+import de.juplo.kafka.wordcount.counter.TestOutputUser;
+import de.juplo.kafka.wordcount.counter.TestOutputWord;
+import de.juplo.kafka.wordcount.recorder.TestInputRecording;
+import de.juplo.kafka.wordcount.recorder.TestInputUser;
import lombok.extern.slf4j.Slf4j;
import org.junit.jupiter.api.BeforeAll;
import org.junit.jupiter.api.Test;
import org.springframework.util.MultiValueMap;
import java.time.Duration;
-import java.util.stream.Stream;
import static de.juplo.kafka.wordcount.splitter.SplitterApplicationIT.TOPIC_IN;
import static de.juplo.kafka.wordcount.splitter.SplitterApplicationIT.TOPIC_OUT;
@SpringBootTest(
properties = {
+ "spring.kafka.producer.key-serializer=org.springframework.kafka.support.serializer.JsonSerializer",
"spring.kafka.producer.value-serializer=org.springframework.kafka.support.serializer.JsonSerializer",
"spring.kafka.producer.properties.spring.json.add.type.headers=false",
"spring.kafka.consumer.auto-offset-reset=earliest",
+ "spring.kafka.consumer.key-deserializer=org.springframework.kafka.support.serializer.JsonDeserializer",
"spring.kafka.consumer.value-deserializer=org.springframework.kafka.support.serializer.JsonDeserializer",
- "spring.kafka.consumer.properties.spring.json.value.default.type=de.juplo.kafka.wordcount.counter.TestWord",
- "spring.kafka.consumer.properties.spring.json.trusted.packages=de.juplo.kafka.wordcount.splitter",
+ "spring.kafka.consumer.properties.spring.json.key.default.type=de.juplo.kafka.wordcount.counter.TestOutputUser",
+ "spring.kafka.consumer.properties.spring.json.value.default.type=de.juplo.kafka.wordcount.counter.TestOutputWord",
"logging.level.root=WARN",
"logging.level.de.juplo=DEBUG",
"juplo.wordcount.splitter.bootstrap-server=${spring.embedded.kafka.brokers}",
@BeforeAll
public static void testSendMessage(
- @Autowired KafkaTemplate<String, TestRecording> kafkaTemplate)
+ @Autowired KafkaTemplate<TestInputUser, TestInputRecording> kafkaTemplate)
{
- Stream
- .of(TestData.INPUT_MESSAGES)
+ TestData
+ .getInputMessages()
.forEach(kv ->
{
try
{
- SendResult<String, TestRecording> result = kafkaTemplate.send(TOPIC_IN, kv.key, kv.value).get();
+ SendResult<TestInputUser, TestInputRecording> result = kafkaTemplate.send(TOPIC_IN, kv.key, kv.value).get();
log.info(
"Sent: {}={}, partition={}, offset={}",
result.getProducerRecord().key(),
@Test
void testSendMessage() throws Exception
{
- await("Expexted converted data")
+ await("Expected converted data")
.atMost(Duration.ofSeconds(5))
.untilAsserted(() ->
TestData.assertExpectedMessages(consumer.getReceivedMessages()));
static class Consumer
{
- private final MultiValueMap<String, TestWord> received = new LinkedMultiValueMap<>();
+ private final MultiValueMap<TestOutputUser, TestOutputWord> received = new LinkedMultiValueMap<>();
@KafkaListener(groupId = "TEST", topics = TOPIC_OUT)
public synchronized void receive(
- @Header(KafkaHeaders.RECEIVED_KEY) String key,
- @Payload TestWord value)
+ @Header(KafkaHeaders.RECEIVED_KEY) TestOutputUser key,
+ @Payload TestOutputWord value)
{
log.debug("Received message: {}={}", key, value);
received.add(key, value);
}
- synchronized MultiValueMap<String, TestWord> getReceivedMessages()
+ synchronized MultiValueMap<TestOutputUser, TestOutputWord> getReceivedMessages()
{
return received;
}