Set<ConsumerRecord<String, Long>> received = new HashSet<>();
testHandler = record -> received.add(record);
+ endlessConsumer.start();
+
await("100 records received")
.atMost(Duration.ofSeconds(30))
.until(() -> received.size() >= 100);
? new Bytes(stringSerializer.serialize(TOPIC, "BOOM!"))
: new Bytes(longSerializer.serialize(TOPIC, counter)));
+ Set<ConsumerRecord<String, Long>> received = new HashSet<>();
+ testHandler = record -> received.add(record);
+
+ endlessConsumer.start();
+
await("Consumer failed")
.atMost(Duration.ofSeconds(30))
.until(() -> !endlessConsumer.running());
checkSeenOffsetsForProgress();
compareToCommitedOffsets(newOffsets);
+ assertThat(received.size())
+ .describedAs("Received not all sent events")
+ .isLessThan(100);
}
{
Long expected = offsetsToCheck.get(tp) + 1;
log.debug("Checking, if the offset for {} is {}", tp, expected);
- assertThat(offset).isEqualTo(expected);
+ assertThat(offset)
+ .describedAs("Committed offset corresponds to the offset of the consumer")
+ .isEqualTo(expected);
});
}
withProgress.add(tp);
}
});
- assertThat(withProgress).isNotEmpty().describedAs("Found no partitions with any offset-progress");
+ assertThat(withProgress)
+ .describedAs("Some offsets must have changed, compared to the old offset-positions")
+ .isNotEmpty();
}
properties.getTopic(),
kafkaConsumer,
captureOffsetAndExecuteTestHandler);
-
- endlessConsumer.start();
}
@AfterEach