NEU
[demos/kafka/chat] / src / main / java / de / juplo / kafka / chat / backend / persistence / kafka / KafkaChatHomeService.java
index e171bc5..f105902 100644 (file)
@@ -2,57 +2,81 @@ package de.juplo.kafka.chat.backend.persistence.kafka;
 
 import de.juplo.kafka.chat.backend.domain.ChatHomeService;
 import de.juplo.kafka.chat.backend.domain.ChatRoom;
+import de.juplo.kafka.chat.backend.domain.Message;
+import de.juplo.kafka.chat.backend.domain.ShardNotOwnedException;
+import de.juplo.kafka.chat.backend.persistence.KafkaLikeShardingStrategy;
 import lombok.extern.slf4j.Slf4j;
 import org.apache.kafka.clients.consumer.Consumer;
 import org.apache.kafka.clients.consumer.ConsumerRebalanceListener;
+import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.clients.consumer.ConsumerRecords;
+import org.apache.kafka.clients.producer.Producer;
+import org.apache.kafka.clients.producer.ProducerRecord;
 import org.apache.kafka.common.TopicPartition;
 import reactor.core.publisher.Flux;
 import reactor.core.publisher.Mono;
 
+import java.time.*;
 import java.util.*;
+import java.util.concurrent.ExecutorService;
 
 
 @Slf4j
-public class KafkaChatHomeService implements ChatHomeService, ConsumerRebalanceListener
+public class KafkaChatHomeService implements ChatHomeService, Runnable, ConsumerRebalanceListener
 {
+  private final ExecutorService executorService;
   private final Consumer<String, MessageTo> consumer;
+  private final Producer<String, MessageTo> producer;
   private final String topic;
+  private final ZoneId zoneId;
   // private final long[] offsets; Erst mal immer alles neu einlesen
-  private final Map<UUID, KafkaChatRoomService>[] kafkaChatRoomServiceMaps;
+  private final boolean[] isShardOwned;
   private final Map<UUID, ChatRoom>[] chatRoomMaps;
+  private final KafkaLikeShardingStrategy shardingStrategy;
+
+  private boolean running;
+  private volatile boolean loadInProgress;
 
 
   public KafkaChatHomeService(
+    ExecutorService executorService,
     Consumer<String, MessageTo> consumer,
+    Producer<String, MessageTo> producer,
     String topic,
+    ZoneId zoneId,
     int numShards)
   {
     log.debug("Creating KafkaChatHomeService");
+    this.executorService = executorService;
     this.consumer = consumer;
+    this.producer = producer;
     this.topic = topic;
+    this.zoneId = zoneId;
     // this.offsets = new long[numShards];
     // for (int i=0; i< numShards; i++)
     // {
     //   this.offsets[i] = 0l;
     // }
-    this.kafkaChatRoomServiceMaps = new Map[numShards];
+    this.isShardOwned = new boolean[numShards];
     this.chatRoomMaps = new Map[numShards];
+    this.shardingStrategy = new KafkaLikeShardingStrategy(numShards);
   }
 
 
   @Override
   public void onPartitionsAssigned(Collection<TopicPartition> partitions)
   {
-    consumer.endOffsets(partitions).forEach((tp, currentOffset) ->
+    loadInProgress = true;
+
+    consumer.endOffsets(partitions).forEach((topicPartition, currentOffset) ->
     {
-      if (!tp.topic().equals(topic))
+      if (!topicPartition.topic().equals(topic))
       {
-        log.warn("Ignoring partition from unwanted topic: {}", tp);
+        log.warn("Ignoring partition from unwanted topic: {}", topicPartition);
         return;
       }
 
-      int partition = tp.partition();
-      kafkaChatRoomServiceMaps[partition] = new HashMap<>(); // TODO: reuse! Nicht immer alles neu laden
+      int partition = topicPartition.partition();
       long unseenOffset = 0; // offsets[partition];
 
       log.info(
@@ -61,29 +85,26 @@ public class KafkaChatHomeService implements ChatHomeService, ConsumerRebalanceL
           unseenOffset,
           currentOffset);
 
-      consumer.seek(tp, unseenOffset);
-      chatRoomMaps[partition]
-          .values()
-          .stream()
-      handlers[partition] = new ChatRoomLoadingMessageHandlingStrategy(tp, currentOffset, unseenOffset);
+      // TODO: reuse! Nicht immer alles neu laden, sondern erst ab offsets[partition]!
+      consumer.seek(topicPartition, unseenOffset);
     });
+
+    consumer.resume(partitions);
   }
 
   @Override
   public void onPartitionsRevoked(Collection<TopicPartition> partitions)
   {
-    partitions.forEach(tp ->
+    partitions.forEach(topicPartition ->
     {
-      if (!tp.topic().equals(topic))
+      if (!topicPartition.topic().equals(topic))
       {
-        log.warn("Ignoring partition from unwanted topic: {}", tp);
+        log.warn("Ignoring partition from unwanted topic: {}", topicPartition);
         return;
       }
 
-      int partition = tp.partition();
-      long unseenOffset = offsets[partition];
-
-      log.info("Reading partition {} from {} -> {}", partition, unseenOffset, currentOffset);
+      int partition = topicPartition.partition();
+      // long unseenOffset = offsets[partition]; TODO: Offset merken...?
     });
     log.info("Revoked partitions: {}", partitions);
   }
@@ -91,56 +112,119 @@ public class KafkaChatHomeService implements ChatHomeService, ConsumerRebalanceL
   @Override
   public void onPartitionsLost(Collection<TopicPartition> partitions)
   {
-    log.info("Revoked partitions: {}", partitions);
+    // TODO: Muss auf den Verlust anders reagiert werden?
+    onPartitionsRevoked(partitions);
   }
 
-  private void foo()
+  @Override
+  public void run()
   {
-    Set<Integer> owned = Arrays
-      .stream(ownedShards)
-      .collect(
-        () -> new HashSet<>(),
-        (set, i) -> set.add(i),
-        (a, b) -> a.addAll(b));
-    for (int shard = 0; shard < numShards; shard++)
+    consumer.subscribe(List.of(topic));
+
+    running = true;
+
+    try
     {
-      chatRoomMaps[shard] = owned.contains(shard)
-        ? new HashMap<>()
-        : null;
-    }
-    chatroomFlux
-      .filter(chatRoom ->
+      while (running)
       {
-        if (owned.contains(chatRoom.getShard()))
+        ConsumerRecords<String, MessageTo> records = consumer.poll(Duration.ofMinutes(5));
+        log.info("Fetched {} messages", records.count());
+
+        if (loadInProgress)
         {
-          return true;
+          for (ConsumerRecord<String, MessageTo> record : records)
+          {
+            UUID chatRoomId = UUID.fromString(record.key());
+            MessageTo messageTo = record.value();
+            ChatRoom chatRoom = chatRoomMaps[record.partition()].get(chatRoomId);
+            KafkaChatRoomService kafkaChatRoomService =
+                (KafkaChatRoomService) chatRoom.getChatRoomService();
+            Message.MessageKey key = Message.MessageKey.of(messageTo.getUser(), messageTo.getId());
+            Instant instant = Instant.ofEpochSecond(record.timestamp());
+            LocalDateTime timestamp = LocalDateTime.ofInstant(instant, zoneId);
+            Message message = new Message(key, record.offset(), timestamp, messageTo.getText());
+            kafkaChatRoomService.persistMessage(message);
+          }
         }
         else
         {
-          log.info("Ignoring not owned chat-room {}", chatRoom);
-          return false;
+          if (!records.isEmpty())
+          {
+            throw new IllegalStateException("All owned partions should be paused, when no load is in progress!");
+          }
         }
-      })
-      .toStream()
-      .forEach(chatroom -> chatRoomMaps[chatroom.getShard()].put(chatroom.getId(), chatroom));
+      }
+    }
   }
 
-  @Override
-  public Mono<ChatRoom> putChatRoom(ChatRoom chatRoom)
+  Mono<Message> sendMessage(
+      UUID chatRoomId,
+      Message.MessageKey key,
+      LocalDateTime timestamp,
+      String text)
   {
-    chatRoomMaps[chatRoom.getShard()].put(chatRoom.getId(), chatRoom);
-    return Mono.just(chatRoom);
+    int shard = this.shardingStrategy.selectShard(chatRoomId);
+    TopicPartition tp = new TopicPartition(topic, shard);
+    ZonedDateTime zdt = ZonedDateTime.of(timestamp, zoneId);
+    return Mono.create(sink ->
+    {
+      ProducerRecord<String, MessageTo> record =
+          new ProducerRecord<>(
+              tp.topic(),
+              tp.partition(),
+              zdt.toEpochSecond(),
+              chatRoomId.toString(),
+              MessageTo.of(key.getUsername(), key.getMessageId(), text));
+
+      producer.send(record, ((metadata, exception) ->
+      {
+        if (metadata != null)
+        {
+          // On successful send
+          Message message = new Message(key, metadata.offset(), timestamp, text);
+          log.info("Successfully send message {}", message);
+          sink.success(message);
+        }
+        else
+        {
+          // On send-failure
+          log.error(
+              "Could not send message for chat-room={}, key={}, timestamp={}, text={}: {}",
+              chatRoomId,
+              key,
+              timestamp,
+              text,
+              exception);
+          sink.error(exception);
+        }
+      }));
+    });
   }
 
+
   @Override
   public Mono<ChatRoom> getChatRoom(int shard, UUID id)
   {
-    return Mono.justOrEmpty(chatRoomMaps[shard].get(id));
+    if (loadInProgress)
+    {
+      throw new ShardNotOwnedException(shard);
+    }
+    else
+    {
+      return Mono.justOrEmpty(chatRoomMaps[shard].get(id));
+    }
   }
 
   @Override
   public Flux<ChatRoom> getChatRooms(int shard)
   {
-    return Flux.fromStream(chatRoomMaps[shard].values().stream());
+    if (loadInProgress)
+    {
+      throw new ShardNotOwnedException(shard);
+    }
+    else
+    {
+      return Flux.fromStream(chatRoomMaps[shard].values().stream());
+    }
   }
 }