@Order(1) // << The poistion pill is not skipped. Hence, this test must run first
void commitsCurrentOffsetsOnSuccess() throws ExecutionException, InterruptedException
{
- send100Messages((key, counter) -> serialize(key, counter));
+ send100Messages((partition, key, counter) ->
+ {
- Bytes value = new Bytes(valueSerializer.serialize(TOPIC, counter));
++ Bytes value = serialize(key, counter);
+ return new ProducerRecord<>(TOPIC, partition, key, value);
+ });
await("100 records received")
.atMost(Duration.ofSeconds(30))
@Order(2)
void commitsOffsetOfErrorForReprocessingOnError()
{
- send100Messages((key, counter) ->
- counter == 77
- ? new Bytes(stringSerializer.serialize(TOPIC, "BOOM!"))
- : serialize(key, counter));
+ send100Messages((partition, key, counter) ->
+ {
+ Bytes value = counter == 77
+ ? new Bytes(stringSerializer.serialize(TOPIC, "BOOM!"))
- : new Bytes(valueSerializer.serialize(TOPIC, counter));
++ : serialize(key, counter);
+ return new ProducerRecord<>(TOPIC, partition, key, value);
+ });
await("Consumer failed")
.atMost(Duration.ofSeconds(30))
{
for (int key = 0; key < 10; key++)
{
- Bytes value = messageGenerator.apply(key, ++i);
-
ProducerRecord<String, Bytes> record =
- new ProducerRecord<>(
- TOPIC,
- partition,
- Integer.toString(key%2),
- value);
+ recordGenerator.generate(partition, Integer.toString(partition*10+key%2), ++i);
+ record.headers().add("__TypeId__", "message".getBytes());
kafkaProducer.send(record, (metadata, e) ->
{
if (metadata != null)
}
}
- Bytes serialize(Integer key, Long value)
++ Bytes serialize(String key, Long value)
+ {
+ ClientMessage message = new ClientMessage();
- message.setClient(key.toString());
++ message.setClient(key);
+ message.setMessage(value.toString());
+ return new Bytes(valueSerializer.serialize(TOPIC, message));
+ }
+
@BeforeEach
public void init()