package de.juplo.kafka.wordcount.counter;
+import de.juplo.kafka.wordcount.splitter.TestInputUser;
import de.juplo.kafka.wordcount.splitter.TestInputWord;
import de.juplo.kafka.wordcount.top10.TestOutputWord;
import de.juplo.kafka.wordcount.top10.TestOutputWordCounter;
import lombok.extern.slf4j.Slf4j;
-import org.apache.kafka.common.serialization.StringSerializer;
import org.apache.kafka.streams.TestInputTopic;
import org.apache.kafka.streams.TestOutputTopic;
import org.apache.kafka.streams.Topology;
import org.springframework.util.MultiValueMap;
import static de.juplo.kafka.wordcount.counter.CounterApplicationConfiguriation.serializationConfig;
+import static de.juplo.kafka.wordcount.counter.CounterStreamProcessor.STORE_NAME;
@Slf4j
{
public static final String IN = "TEST-IN";
public static final String OUT = "TEST-OUT";
- public static final String STORE_NAME = "TOPOLOGY-TEST";
TopologyTestDriver testDriver;
- TestInputTopic<String, TestInputWord> in;
+ TestInputTopic<TestInputUser, TestInputWord> in;
TestOutputTopic<TestOutputWord, TestOutputWordCounter> out;
in = testDriver.createInputTopic(
IN,
- new StringSerializer(),
+ new JsonSerializer().noTypeInfo(),
new JsonSerializer().noTypeInfo());
out = testDriver.createOutputTopic(
OUT,
- new JsonDeserializer()
- .copyWithType(TestOutputWord.class)
- .ignoreTypeHeaders(),
- new JsonDeserializer()
- .copyWithType(TestOutputWordCounter.class)
- .ignoreTypeHeaders());
+ new JsonDeserializer(TestOutputWord.class).ignoreTypeHeaders(),
+ new JsonDeserializer(TestOutputWordCounter.class).ignoreTypeHeaders());
}