X-Git-Url: https://juplo.de/gitweb/?a=blobdiff_plain;f=src%2Fmain%2Fjava%2Fde%2Fjuplo%2Fkafka%2Fwordcount%2Fquery%2FQueryStreamProcessor.java;h=c4ae4ae0a0ecdd3cee579096b16b4006060968c3;hb=2b22a006cc57203406c8589687a6c729ebdbf40c;hp=319861da3d3e6c0ad9cc491c4fd13d1d92d37c56;hpb=3bfe34e8bc0539932cf93361fe6f710738b37897;p=demos%2Fkafka%2Fwordcount diff --git a/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java b/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java index 319861d..c4ae4ae 100644 --- a/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java +++ b/src/main/java/de/juplo/kafka/wordcount/query/QueryStreamProcessor.java @@ -2,80 +2,104 @@ package de.juplo.kafka.wordcount.query; import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.ObjectMapper; +import jakarta.annotation.PostConstruct; +import jakarta.annotation.PreDestroy; import lombok.extern.slf4j.Slf4j; -import org.apache.kafka.clients.consumer.ConsumerConfig; import org.apache.kafka.common.serialization.Serdes; import org.apache.kafka.streams.*; +import org.apache.kafka.streams.kstream.KStream; +import org.apache.kafka.streams.kstream.KTable; import org.apache.kafka.streams.kstream.Materialized; import org.apache.kafka.streams.state.HostInfo; +import org.apache.kafka.streams.state.KeyValueBytesStoreSupplier; import org.apache.kafka.streams.state.QueryableStoreTypes; import org.apache.kafka.streams.state.ReadOnlyKeyValueStore; -import org.springframework.boot.SpringApplication; -import org.springframework.context.ConfigurableApplicationContext; -import org.springframework.stereotype.Component; -import javax.annotation.PostConstruct; -import javax.annotation.PreDestroy; import java.net.URI; import java.util.Optional; import java.util.Properties; -import java.util.concurrent.CompletableFuture; - -import static org.apache.kafka.streams.errors.StreamsUncaughtExceptionHandler.StreamThreadExceptionResponse.SHUTDOWN_CLIENT; @Slf4j -@Component public class QueryStreamProcessor { + public static final String STORE_NAME = "rankings-by-username"; + public final KafkaStreams streams; public final HostInfo hostInfo; - public final String storeName = "rankingsByUsername"; public final StoreQueryParameters> storeParameters; public final ObjectMapper mapper; public QueryStreamProcessor( - QueryApplicationProperties properties, - ObjectMapper mapper, - ConfigurableApplicationContext context) + Properties props, + HostInfo applicationServer, + String usersInputTopic, + String rankingInputTopic, + KeyValueBytesStoreSupplier storeSupplier, + ObjectMapper mapper) { - StreamsBuilder builder = new StreamsBuilder(); + Topology topology = buildTopology( + usersInputTopic, + rankingInputTopic, + storeSupplier, + mapper); + streams = new KafkaStreams(topology, props); + hostInfo = applicationServer; + storeParameters = StoreQueryParameters.fromNameAndType(STORE_NAME, QueryableStoreTypes.keyValueStore());; + this.mapper = mapper; + } - builder.table(properties.getRankingInputTopic(), Materialized.as(storeName)); + static Topology buildTopology( + String usersInputTopic, + String rankingInputTopic, + KeyValueBytesStoreSupplier storeSupplier, + ObjectMapper mapper) + { + StreamsBuilder builder = new StreamsBuilder(); - Properties props = new Properties(); - props.put(StreamsConfig.APPLICATION_ID_CONFIG, properties.getApplicationId()); - props.put(StreamsConfig.APPLICATION_SERVER_CONFIG, properties.getApplicationServer()); - props.put(StreamsConfig.BOOTSTRAP_SERVERS_CONFIG, properties.getBootstrapServer()); - props.put(StreamsConfig.DEFAULT_KEY_SERDE_CLASS_CONFIG, Serdes.String().getClass().getName()); - props.put(StreamsConfig.DEFAULT_VALUE_SERDE_CLASS_CONFIG, Serdes.String().getClass().getName()); - props.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); + KTable users = builder.table(usersInputTopic); + KStream rankings = builder.stream(rankingInputTopic); + + rankings + .join(users, (rankingJson, userJson) -> + { + try + { + Ranking ranking = mapper.readValue(rankingJson, Ranking.class); + User user = mapper.readValue(userJson, User.class); + + return mapper.writeValueAsString( + UserRanking.of( + user.getFirstName(), + user.getLastName(), + ranking.getEntries())); + } + catch (JsonProcessingException e) + { + throw new RuntimeException(e); + } + }) + .toTable(Materialized.as(storeSupplier)); + + Topology topology = builder.build(); + log.info("\n\n{}", topology.describe()); + + return topology; + } - streams = new KafkaStreams(builder.build(), props); - streams.setUncaughtExceptionHandler((Throwable e) -> - { - log.error("Unexpected error!", e); - CompletableFuture.runAsync(() -> - { - log.info("Stopping application..."); - SpringApplication.exit(context, () -> 1); - }); - return SHUTDOWN_CLIENT; - }); - - hostInfo = HostInfo.buildFromEndpoint(properties.getApplicationServer()); - storeParameters = StoreQueryParameters.fromNameAndType(storeName, QueryableStoreTypes.keyValueStore());; - this.mapper = mapper; + ReadOnlyKeyValueStore getStore() + { + return streams.store(storeParameters); } public Optional getRedirect(String username) { - KeyQueryMetadata metadata = streams.queryMetadataForKey(storeName, username, Serdes.String().serializer()); + KeyQueryMetadata metadata = streams.queryMetadataForKey(STORE_NAME, username, Serdes.String().serializer()); HostInfo activeHost = metadata.activeHost(); log.debug("Local store for {}: {}, {}:{}", username, metadata.partition(), activeHost.host(), activeHost.port()); - if (activeHost.equals(this.hostInfo)) + if (activeHost.equals(this.hostInfo) || activeHost.equals(HostInfo.unavailable())) { return Optional.empty(); } @@ -85,16 +109,16 @@ public class QueryStreamProcessor return Optional.of(location); } - public Optional getRanking(String username) + public Optional getUserRanking(String username) { return Optional - .ofNullable(streams.store(storeParameters).get(username)) + .ofNullable(getStore().get(username)) .map(json -> { try { - return mapper.readValue(json, Ranking.class); + return mapper.readValue(json, UserRanking.class); } catch (JsonProcessingException e) {