X-Git-Url: https://juplo.de/gitweb/?a=blobdiff_plain;f=src%2Fmain%2Fjava%2Fde%2Fjuplo%2Fkafka%2Fwordcount%2Fquery%2FQueryStreamProcessor.java;h=ff7c1501c0a7cae4e2373d729bf240f18677ff47;hb=57f47fca712981116b726e437f589ac727c5d0a7;hp=696e08862abd823f6508cd1e0e18d4227b8a15c4;hpb=0d1ed8c65f422e5c0cd291c35dba67b1f2205997;p=demos%2Fkafka%2Fwordcount diff --git a/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java b/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java index 696e088..ff7c150 100644 --- a/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java +++ b/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java @@ -2,50 +2,64 @@ package de.juplo.kafka.wordcount.query; import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.ObjectMapper; +import jakarta.annotation.PostConstruct; +import jakarta.annotation.PreDestroy; import lombok.extern.slf4j.Slf4j; -import org.apache.kafka.clients.consumer.ConsumerConfig; import org.apache.kafka.common.serialization.Serdes; import org.apache.kafka.streams.*; import org.apache.kafka.streams.kstream.KStream; import org.apache.kafka.streams.kstream.KTable; import org.apache.kafka.streams.kstream.Materialized; import org.apache.kafka.streams.state.HostInfo; +import org.apache.kafka.streams.state.KeyValueBytesStoreSupplier; import org.apache.kafka.streams.state.QueryableStoreTypes; import org.apache.kafka.streams.state.ReadOnlyKeyValueStore; -import org.springframework.boot.SpringApplication; -import org.springframework.context.ConfigurableApplicationContext; -import org.springframework.stereotype.Component; -import javax.annotation.PostConstruct; -import javax.annotation.PreDestroy; import java.net.URI; import java.util.Optional; import java.util.Properties; -import java.util.concurrent.CompletableFuture; - -import static org.apache.kafka.streams.errors.StreamsUncaughtExceptionHandler.StreamThreadExceptionResponse.SHUTDOWN_CLIENT; @Slf4j -@Component public class QueryStreamProcessor { + public static final String STORE_NAME = "rankings-by-username"; + public final KafkaStreams streams; public final HostInfo hostInfo; - public final String storeName = "rankingsByUsername"; public final StoreQueryParameters> storeParameters; public final ObjectMapper mapper; public QueryStreamProcessor( - QueryApplicationProperties properties, - ObjectMapper mapper, - ConfigurableApplicationContext context) + Properties props, + HostInfo applicationServer, + String usersInputTopic, + String rankingInputTopic, + KeyValueBytesStoreSupplier storeSupplier, + ObjectMapper mapper) + { + Topology topology = buildTopology( + usersInputTopic, + rankingInputTopic, + storeSupplier, + mapper); + streams = new KafkaStreams(topology, props); + hostInfo = applicationServer; + storeParameters = StoreQueryParameters.fromNameAndType(STORE_NAME, QueryableStoreTypes.keyValueStore());; + this.mapper = mapper; + } + + static Topology buildTopology( + String usersInputTopic, + String rankingInputTopic, + KeyValueBytesStoreSupplier storeSupplier, + ObjectMapper mapper) { StreamsBuilder builder = new StreamsBuilder(); - KTable users = builder.table(properties.getUsersInputTopic()); - KStream rankings = builder.stream(properties.getRankingInputTopic()); + KTable users = builder.table(usersInputTopic); + KStream rankings = builder.stream(rankingInputTopic); rankings .join(users, (rankingJson, userJson) -> @@ -66,36 +80,22 @@ public class QueryStreamProcessor throw new RuntimeException(e); } }) - .toTable(Materialized.as(storeName)); - - Properties props = new Properties(); - props.put(StreamsConfig.APPLICATION_ID_CONFIG, properties.getApplicationId()); - props.put(StreamsConfig.APPLICATION_SERVER_CONFIG, properties.getApplicationServer()); - props.put(StreamsConfig.BOOTSTRAP_SERVERS_CONFIG, properties.getBootstrapServer()); - props.put(StreamsConfig.DEFAULT_KEY_SERDE_CLASS_CONFIG, Serdes.String().getClass().getName()); - props.put(StreamsConfig.DEFAULT_VALUE_SERDE_CLASS_CONFIG, Serdes.String().getClass().getName()); - props.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); - - streams = new KafkaStreams(builder.build(), props); - streams.setUncaughtExceptionHandler((Throwable e) -> - { - log.error("Unexpected error!", e); - CompletableFuture.runAsync(() -> - { - log.info("Stopping application..."); - SpringApplication.exit(context, () -> 1); - }); - return SHUTDOWN_CLIENT; - }); - - hostInfo = HostInfo.buildFromEndpoint(properties.getApplicationServer()); - storeParameters = StoreQueryParameters.fromNameAndType(storeName, QueryableStoreTypes.keyValueStore());; - this.mapper = mapper; + .toTable(Materialized.as(storeSupplier)); + + Topology topology = builder.build(); + log.info("\n\n{}", topology.describe()); + + return topology; + } + + ReadOnlyKeyValueStore getStore() + { + return streams.store(storeParameters); } public Optional getRedirect(String username) { - KeyQueryMetadata metadata = streams.queryMetadataForKey(storeName, username, Serdes.String().serializer()); + KeyQueryMetadata metadata = streams.queryMetadataForKey(STORE_NAME, username, Serdes.String().serializer()); HostInfo activeHost = metadata.activeHost(); log.debug("Local store for {}: {}, {}:{}", username, metadata.partition(), activeHost.host(), activeHost.port()); @@ -113,7 +113,7 @@ public class QueryStreamProcessor { return Optional - .ofNullable(streams.store(storeParameters).get(username)) + .ofNullable(getStore().get(username)) .map(json -> { try