import org.junit.jupiter.api.*;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.boot.autoconfigure.EnableAutoConfiguration;
-import org.springframework.boot.test.autoconfigure.data.mongo.AutoConfigureDataMongo;
import org.springframework.boot.test.context.ConfigDataApplicationContextInitializer;
import org.springframework.boot.test.context.TestConfiguration;
import org.springframework.context.annotation.Bean;
properties = {
"consumer.bootstrap-server=${spring.embedded.kafka.brokers}",
"consumer.topic=" + TOPIC,
- "consumer.commit-interval=1s",
- "spring.mongodb.embedded.version=4.4.13" })
+ "consumer.commit-interval=1s" })
@EmbeddedKafka(topics = TOPIC, partitions = PARTITIONS)
@EnableAutoConfiguration
-@AutoConfigureDataMongo
@Slf4j
class ApplicationTests
{
@Autowired
ExecutorService executor;
@Autowired
- KeyCountingRebalanceListener keyCountingRebalanceListener;
- @Autowired
- KeyCountingRecordHandler keyCountingRecordHandler;
+ RecordHandler noopRecordHandler;
EndlessConsumer<String, Long> endlessConsumer;
Map<TopicPartition, Long> oldOffsets;
});
TestRecordHandler<String, Long> captureOffsetAndExecuteTestHandler =
- new TestRecordHandler<String, Long>(keyCountingRecordHandler) {
+ new TestRecordHandler<String, Long>(noopRecordHandler) {
@Override
public void onNewRecord(ConsumerRecord<String, Long> record)
{
properties.getClientId(),
properties.getTopic(),
kafkaConsumer,
- keyCountingRebalanceListener,
captureOffsetAndExecuteTestHandler);
endlessConsumer.start();