projects
/
demos
/
kafka
/
training
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Konfig-Parameter zum künstlichen Verzögern der Verabeitung eingebaut
[demos/kafka/training]
/
src
/
main
/
java
/
de
/
juplo
/
kafka
/
ApplicationConfiguration.java
diff --git
a/src/main/java/de/juplo/kafka/ApplicationConfiguration.java
b/src/main/java/de/juplo/kafka/ApplicationConfiguration.java
index
8e2e867
..
4d056c4
100644
(file)
--- a/
src/main/java/de/juplo/kafka/ApplicationConfiguration.java
+++ b/
src/main/java/de/juplo/kafka/ApplicationConfiguration.java
@@
-1,14
+1,13
@@
package de.juplo.kafka;
package de.juplo.kafka;
-import org.apache.kafka.clients.consumer.Consumer;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.clients.consumer.KafkaConsumer;
-import org.apache.kafka.common.serialization.LongDeserializer;
import org.apache.kafka.common.serialization.StringDeserializer;
import org.springframework.boot.context.properties.EnableConfigurationProperties;
import org.springframework.context.annotation.Bean;
import org.springframework.context.annotation.Configuration;
import java.time.Clock;
import org.apache.kafka.common.serialization.StringDeserializer;
import org.springframework.boot.context.properties.EnableConfigurationProperties;
import org.springframework.context.annotation.Bean;
import org.springframework.context.annotation.Configuration;
import java.time.Clock;
+import java.util.Optional;
import java.util.Properties;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.Properties;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
@@
-19,34
+18,41
@@
import java.util.concurrent.Executors;
public class ApplicationConfiguration
{
@Bean
public class ApplicationConfiguration
{
@Bean
- public KeyCountingRecordHandler messageCountingRecordHandler()
+ public ApplicationRecordHandler recordHandler(
+ AdderResults adderResults,
+ ApplicationProperties properties)
+ {
+ return new ApplicationRecordHandler(
+ adderResults,
+ Optional.ofNullable(properties.getThrottle()));
+ }
+
+ @Bean
+ public AdderResults adderResults()
{
{
- return new
KeyCountingRecordHandler
();
+ return new
AdderResults
();
}
@Bean
}
@Bean
- public
KeyCountingRebalanceListener wordcountR
ebalanceListener(
-
KeyCountingRecordHandler keyCountingR
ecordHandler,
-
PartitionStatisticsRepository repository
,
-
Consumer<String, Long> consumer
,
+ public
ApplicationRebalanceListener r
ebalanceListener(
+
ApplicationRecordHandler r
ecordHandler,
+
AdderResults adderResults
,
+
StateRepository stateRepository
,
ApplicationProperties properties)
{
ApplicationProperties properties)
{
- return new KeyCountingRebalanceListener(
- keyCountingRecordHandler,
- repository,
- properties.getClientId(),
- properties.getTopic(),
- Clock.systemDefaultZone(),
- properties.getCommitInterval(),
- consumer);
+ return new ApplicationRebalanceListener(
+ recordHandler,
+ adderResults,
+ stateRepository,
+ properties.getClientId());
}
@Bean
}
@Bean
- public EndlessConsumer<String,
Lo
ng> endlessConsumer(
- KafkaConsumer<String,
Lo
ng> kafkaConsumer,
+ public EndlessConsumer<String,
Stri
ng> endlessConsumer(
+ KafkaConsumer<String,
Stri
ng> kafkaConsumer,
ExecutorService executor,
ExecutorService executor,
-
KeyCountingRebalanceListener keyCountingR
ebalanceListener,
-
KeyCountingRecordHandler keyCountingR
ecordHandler,
+
ApplicationRebalanceListener r
ebalanceListener,
+
ApplicationRecordHandler r
ecordHandler,
ApplicationProperties properties)
{
return
ApplicationProperties properties)
{
return
@@
-55,8
+61,8
@@
public class ApplicationConfiguration
properties.getClientId(),
properties.getTopic(),
kafkaConsumer,
properties.getClientId(),
properties.getTopic(),
kafkaConsumer,
-
keyCountingR
ebalanceListener,
-
keyCountingR
ecordHandler);
+
r
ebalanceListener,
+
r
ecordHandler);
}
@Bean
}
@Bean
@@
-66,7
+72,7
@@
public class ApplicationConfiguration
}
@Bean(destroyMethod = "close")
}
@Bean(destroyMethod = "close")
- public KafkaConsumer<String,
Lo
ng> kafkaConsumer(ApplicationProperties properties)
+ public KafkaConsumer<String,
Stri
ng> kafkaConsumer(ApplicationProperties properties)
{
Properties props = new Properties();
{
Properties props = new Properties();
@@
-74,11
+80,11
@@
public class ApplicationConfiguration
props.put("partition.assignment.strategy", "org.apache.kafka.clients.consumer.CooperativeStickyAssignor");
props.put("group.id", properties.getGroupId());
props.put("client.id", properties.getClientId());
props.put("partition.assignment.strategy", "org.apache.kafka.clients.consumer.CooperativeStickyAssignor");
props.put("group.id", properties.getGroupId());
props.put("client.id", properties.getClientId());
- props.put("enable.auto.commit", false);
props.put("auto.offset.reset", properties.getAutoOffsetReset());
props.put("auto.offset.reset", properties.getAutoOffsetReset());
+ props.put("auto.commit.interval.ms", (int)properties.getCommitInterval().toMillis());
props.put("metadata.max.age.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
props.put("metadata.max.age.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
- props.put("value.deserializer",
Lo
ngDeserializer.class.getName());
+ props.put("value.deserializer",
Stri
ngDeserializer.class.getName());
return new KafkaConsumer<>(props);
}
return new KafkaConsumer<>(props);
}