import java.util.List;
import java.util.Map;
import java.util.Properties;
-import java.util.stream.Collectors;
+
+import static de.juplo.kafka.wordcount.counter.TestData.convertToMap;
public class CounterStreamProcessorTopologyTest
new CounterApplicationConfiguriation();
Properties streamProcessorProperties =
applicationConfiguriation.streamProcessorProperties(new CounterApplicationProperties());
+ Map<String, Object> propertyMap = convertToMap(streamProcessorProperties);
- Map<String, ?> propertyMap = streamProcessorProperties
- .entrySet()
- .stream()
- .collect(
- Collectors.toMap(
- entry -> (String)entry.getKey(),
- entry -> entry.getValue()
- ));
JsonSerde<?> keySerde = new JsonSerde<>();
keySerde.configure(propertyMap, true);
JsonSerde<?> valueSerde = new JsonSerde<>();
(JsonSerializer<String>)keySerde.serializer(),
(JsonSerializer<Word>)valueSerde.serializer());
- TestOutputTopic<Word, WordCount> out = testDriver.createOutputTopic(
+ TestOutputTopic<Word, WordCounter> out = testDriver.createOutputTopic(
OUT,
(JsonDeserializer<Word>)keySerde.deserializer(),
- (JsonDeserializer<WordCount>)valueSerde.deserializer());
+ (JsonDeserializer<WordCounter>)valueSerde.deserializer());
TestData.writeInputData((key, value) -> in.pipeInput(key, value));
- List<KeyValue<Word,WordCount>> receivedMessages = out
+ List<KeyValue<Word, WordCounter>> receivedMessages = out
.readRecordsToList()
.stream()
.map(record -> KeyValue.pair(record.key(), record.value()))