package de.juplo.kafka;
import lombok.extern.slf4j.Slf4j;
-import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.ProducerRecord;
-import org.apache.kafka.common.serialization.StringSerializer;
+import org.apache.kafka.clients.producer.RecordMetadata;
+import org.springframework.kafka.core.KafkaTemplate;
+import org.springframework.kafka.support.SendResult;
+import org.springframework.util.concurrent.ListenableFuture;
import javax.annotation.PreDestroy;
-import java.util.Properties;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.ExecutorService;
-import java.util.concurrent.Future;
@Slf4j
{
private final ExecutorService executor;
private final String id;
- private final String topic;
- private final String acks;
private final int throttleMs;
- private final KafkaProducer<String, String> producer;
+ private final KafkaTemplate<String, String> kafkaTemplate;
private boolean running = false;
private long i = 0;
private long produced = 0;
- private Future<?> future = null;
public EndlessProducer(
ExecutorService executor,
- String bootstrapServer,
String clientId,
- String topic,
- String acks,
- int throttleMs)
+ int throttleMs,
+ KafkaTemplate<String, String> kafkaTemplate)
{
this.executor = executor;
this.id = clientId;
- this.topic = topic;
- this.acks = acks;
this.throttleMs = throttleMs;
-
- Properties props = new Properties();
- props.put("bootstrap.servers", bootstrapServer);
- props.put("client.id", clientId);
- props.put("acks", acks);
- props.put("key.serializer", StringSerializer.class.getName());
- props.put("value.serializer", StringSerializer.class.getName());
-
- this.producer = new KafkaProducer<>(props);
+ this.kafkaTemplate = kafkaTemplate;
}
@Override
{
for (; running; i++)
{
- final long time = System.currentTimeMillis();
-
- final ProducerRecord<String, String> record = new ProducerRecord<>(
- topic, // Topic
- Long.toString(i % 10), // Key
- Long.toString(i) // Value
- );
-
- producer.send(record, (metadata, e) ->
- {
- long now = System.currentTimeMillis();
- if (e == null)
- {
- // HANDLE SUCCESS
- produced++;
- log.debug(
- "{} - Sent key={} message={} partition={}/{} timestamp={} latency={}ms",
- id,
- record.key(),
- record.value(),
- metadata.partition(),
- metadata.offset(),
- metadata.timestamp(),
- now - time
- );
- }
- else
- {
- // HANDLE ERROR
- log.error(
- "{} - ERROR key={} timestamp={} latency={}ms: {}",
- id,
- record.key(),
- metadata == null ? -1 : metadata.timestamp(),
- now - time,
- e.toString()
- );
- }
- });
-
- long now = System.currentTimeMillis();
- log.trace(
- "{} - Queued #{} key={} latency={}ms",
- id,
- i,
- record.key(),
- now - time
- );
+ send(Long.toString(i%10), Long.toString(i));
if (throttleMs > 0)
{
}
catch (Exception e)
{
-
+ log.error("{} - Unexpected Exception:", id, e);
+ }
+ finally
+ {
+ synchronized (this)
+ {
+ running = false;
+ log.info("{} - Stopped - produced {} messages so far", id, produced);
+ }
}
}
+ void send(String key, String value)
+ {
+ final long time = System.currentTimeMillis();
+
+ ListenableFuture<SendResult<String, String>> listenableFuture = kafkaTemplate.sendDefault(key, value);
+ listenableFuture.addCallback(
+ result ->
+ {
+ long now = System.currentTimeMillis();
+ RecordMetadata metadata = result.getRecordMetadata();
+ ProducerRecord<String, String> record = result.getProducerRecord();
+
+ // HANDLE SUCCESS
+ produced++;
+ log.debug(
+ "{} - Sent key={} message={} partition={}/{} timestamp={} latency={}ms",
+ id,
+ record.key(),
+ record.value(),
+ metadata.partition(),
+ metadata.offset(),
+ metadata.timestamp(),
+ now - time
+ );
+ },
+ e ->
+ {
+ long now = System.currentTimeMillis();
+
+ // HANDLE ERROR
+ log.error(
+ "{} - ERROR key={} latency={}ms: {}",
+ id,
+ key,
+ now - time,
+ e.toString()
+ );
+ });
+
+ long now = System.currentTimeMillis();
+ log.trace(
+ "{} - Queued #{} key={} latency={}ms",
+ id,
+ value,
+ key,
+ now - time
+ );
+ }
+
public synchronized void start()
{
if (running)
log.info("{} - Starting - produced {} messages before", id, produced);
running = true;
- future = executor.submit(this);
+ executor.submit(this);
}
public synchronized void stop() throws ExecutionException, InterruptedException
log.info("{} - Stopping...", id);
running = false;
- future.get();
- log.info("{} - Stopped - produced {} messages so far", id, produced);
}
@PreDestroy
}
finally
{
- log.info("{} - Closing the KafkaProducer", id);
- producer.close();
log.info("{}: Produced {} messages in total, exiting!", id, produced);
}
}