All Downloads are FREE. Search and download functionalities are using the official Maven repository.

fs2.kafka.consumer.KafkaConsume.scala Maven / Gradle / Ivy

/*
 * Copyright 2018-2024 OVO Energy Limited
 *
 * SPDX-License-Identifier: Apache-2.0
 */

package fs2.kafka.consumer

import fs2.*
import fs2.kafka.CommittableConsumerRecord

import org.apache.kafka.common.TopicPartition

trait KafkaConsume[F[_], K, V] {

  /**
    * Consume from all assigned partitions, producing a stream of [[CommittableConsumerRecord]]s.
    * Alias for [[stream]].
    */
  final def records: Stream[F, CommittableConsumerRecord[F, K, V]] = stream

  /**
    * Alias for `partitionedStream.parJoinUnbounded`.
    *
    * @note
    *   you have to first use `subscribe` or `assign` the consumer before using this `Stream`. If
    *   you forgot to subscribe, there will be a [[NotSubscribedException]] raised in the `Stream`.
    *
    * @see
    *   [[partitionedRecords]] for more information
    */
  def stream: Stream[F, CommittableConsumerRecord[F, K, V]]

  /**
    * Alias for [[partitionedStream]]
    */
  final def partitionedRecords: Stream[F, Stream[F, CommittableConsumerRecord[F, K, V]]] =
    partitionedStream

  /**
    * `Stream` where the elements themselves are `Stream`s which continually request records for a
    * single partition. These `Stream`s will have to be processed in parallel, using `parJoin` or
    * `parJoinUnbounded`. Note that when using `parJoin(n)` and `n` is smaller than the number of
    * currently assigned partitions, then there will be assigned partitions which won't be
    * processed. For that reason, prefer `parJoinUnbounded` and the actual limit will be the number
    * of assigned partitions.

* * If you do not want to process all partitions in parallel, then you can use [[records]] * instead, where records for all partitions are in a single `Stream`. * * @note * you have to first use `subscribe` or `assign` the consumer before using this `Stream`. If * you forgot to subscribe, there will be a [[NotSubscribedException]] raised in the `Stream`. */ def partitionedStream: Stream[F, Stream[F, CommittableConsumerRecord[F, K, V]]] /** * `Stream` where each element contains a `Map` with all newly assigned partitions. Keys of this * `Map` are `TopicPartition`s, and values are record streams for the particular * `TopicPartition`. These streams will be closed only when a partition is revoked.

* * With the default assignor, all previous partitions are revoked at once, and a new set of * partitions is assigned to a consumer on each rebalance. In this case, each returned `Map` * contains the full partition assignment for the consumer. And all streams from the previous * assignment are closed. It means, that `partitionsMapStream` reflects the default assignment * process in a streaming manner.

* * This may not be the case when a custom assignor is configured in the consumer. When using the * `CooperativeStickyAssignor`, for instance, partitions may be revoked individually. In this * case, each element in the stream (each`Map`) will contain only streams for newly assigned * partitions. Previously returned streams for partitions that are retained will remain active. * Only streams for revoked partitions will be closed.

* * This is the most generic `Stream` method. If you don't need such control, consider using * `partitionedStream` or `stream` methods. They are both based on a `partitionsMapStream`. * * @note * you have to first use `subscribe` or `assign` to subscribe the consumer before using this * `Stream`. If you forgot to subscribe, there will be a [[NotSubscribedException]] raised in * the `Stream`. * @see * [[records]] * @see * [[partitionedRecords]] */ def partitionsMapStream : Stream[F, Map[TopicPartition, Stream[F, CommittableConsumerRecord[F, K, V]]]] /** * Stops consuming new messages from Kafka. This method could be used to implement a graceful * shutdown.

* * This method has a few effects: * - After this call no more data will be fetched from Kafka through the `poll` method. * - All currently running streams will continue to run until all in-flight messages will be * processed. It means that streams will be completed when all fetched messages will be * processed.

* * If some of the [[records]] methods will be called after [[stopConsuming]] call, these methods * will return empty streams.

* * More than one call of [[stopConsuming]] will have no effect. */ def stopConsuming: F[Unit] }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy