fs2.kafka

package fs2.kafka

Members list

Packages

package fs2.kafka.admin

Type members

Classlikes

sealed abstract class Acks

The available options for ProducerSettings#withAcks.

Available options include:
- Acks#Zero to not wait for any acknowledgement from the server,
- Acks#One to only wait for acknowledgement from the leader node,
- Acks#All to wait for acknowledgement from all in-sync replicas
.

The available options for ProducerSettings#withAcks.

Available options include:
- Acks#Zero to not wait for any acknowledgement from the server,
- Acks#One to only wait for acknowledgement from the leader node,
- Acks#All to wait for acknowledgement from all in-sync replicas
.

Attributes

Companion
object
Source
Acks.scala
Supertypes
class Object
trait Matchable
class Any
object Acks

Attributes

Companion
class
Source
Acks.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
Acks.type
sealed abstract class AdminClientSettings

AdminClientSettings contain settings necessary to create a KafkaAdminClient.

AdminClientSettings contain settings necessary to create a KafkaAdminClient. Several convenience functions are provided so that you don't have to work with String values and keys from AdminClientConfig. It's still possible to set AdminClientConfig values with functions like withProperty.

AdminClientSettings instances are immutable and all modification functions return a new AdminClientSettings instance.

Use AdminClientSettings#apply for the default settings, and then apply any desired modifications on top of that instance.

Attributes

Companion
object
Source
AdminClientSettings.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
AdminClientSettings.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class AutoOffsetReset

The available options for ConsumerSettings#withAutoOffsetReset.

Available options include:
- AutoOffsetReset#Earliest to reset to the earliest offsets,
- AutoOffsetReset#Latest to reset to the latest offsets,
- AutoOffsetReset#None to fail if no offsets are available
.

The available options for ConsumerSettings#withAutoOffsetReset.

Available options include:
- AutoOffsetReset#Earliest to reset to the earliest offsets,
- AutoOffsetReset#Latest to reset to the latest offsets,
- AutoOffsetReset#None to fail if no offsets are available
.

Attributes

Companion
object
Source
AutoOffsetReset.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
AutoOffsetReset.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
abstract class CommitRecovery

CommitRecovery describes how to recover from exceptions raised while trying to commit offsets.

CommitRecovery describes how to recover from exceptions raised while trying to commit offsets. See CommitRecovery#Default for the default recovery strategy. If you do not wish to recover from any exceptions, you can use CommitRecovery#None.

To create a new CommitRecovery, simply create a new instance and implement the recoverCommitWith function with the wanted recovery strategy. To use the CommitRecovery, you can simply set it with ConsumerSettings#withCommitRecovery.

Attributes

Companion
object
Source
CommitRecovery.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
CommitRecovery.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class CommitRecoveryException(attempts: Int, lastException: Throwable, offsets: Map[TopicPartition, OffsetAndMetadata]) extends KafkaException

CommitRecoveryException indicates that offset commit recovery was attempted attempts times for offsets, but that it wasn't able to complete successfully.

CommitRecoveryException indicates that offset commit recovery was attempted attempts times for offsets, but that it wasn't able to complete successfully. The last encountered exception is provided as lastException.

Use CommitRecoveryException#apply to create a new instance.

Attributes

Companion
object
Source
CommitRecoveryException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all

Attributes

Companion
class
Source
CommitRecoveryException.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class CommitTimeoutException(timeout: FiniteDuration, offsets: Map[TopicPartition, OffsetAndMetadata]) extends KafkaException

CommitTimeoutException indicates that offset commit took longer than the configured ConsumerSettings#commitTimeout.

CommitTimeoutException indicates that offset commit took longer than the configured ConsumerSettings#commitTimeout. The timeout and offsets are included in the exception message.

Attributes

Source
CommitTimeoutException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class CommittableConsumerRecord[F[_], +K, +V]

CommittableConsumerRecord is a Kafka record along with an instance of CommittableOffset, which can be used commit the record offset to Kafka.

CommittableConsumerRecord is a Kafka record along with an instance of CommittableOffset, which can be used commit the record offset to Kafka. Offsets are normally committed in batches, either using CommittableOffsetBatch or via pipes, like commitBatchWithin. If you are not committing offsets to Kafka then you can use record to get the underlying record and also discard the offset.

While normally not necessary, CommittableConsumerRecord#apply can be used to create a new instance.

Attributes

Companion
object
Source
CommittableConsumerRecord.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
CommittableConsumerRecord.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class CommittableOffset[F[_]]

CommittableOffset represents an offsetAndMetadata for a topicPartition, along with the ability to commit that offset to Kafka with commit.

CommittableOffset represents an offsetAndMetadata for a topicPartition, along with the ability to commit that offset to Kafka with commit. Note that offsets are normally committed in batches for performance reasons. Pipes like commitBatchWithin use CommittableOffsetBatch to commit the offsets in batches.

While normally not necessary, CommittableOffset#apply can be used to create a new instance.

Attributes

Companion
object
Source
CommittableOffset.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
CommittableOffset.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class CommittableOffsetBatch[F[_]]

CommittableOffsetBatch represents a batch of Kafka offsets which can be committed together using commit.

CommittableOffsetBatch represents a batch of Kafka offsets which can be committed together using commit. An offset, or one more batch, can be added an existing batch using updated. Note that this requires the offsets per topic-partition to be included in-order, since offset commits in general require it.

Use CommittableOffsetBatch#empty to create an empty batch. The CommittableOffset#batch function can be used to create a batch from an existing CommittableOffset.

If you have some offsets in-order per topic-partition, you can fold them together using CommittableOffsetBatch#empty and updated, or you can use CommittableOffsetBatch#fromFoldable. Generally, prefer to use fromFoldable, as it has better performance. Provided pipes like commitBatchWithin are also to be preferred, as they also achieve better performance.

Attributes

Companion
object
Source
CommittableOffsetBatch.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
CommittableOffsetBatch.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class CommittableProducerRecords[F[_], +K, +V]

CommittableProducerRecords represents zero or more ProducerRecords and a CommittableOffset, used by TransactionalKafkaProducer to produce the records and commit the offset atomically.

CommittableProducerRecordss can be created using one of the following options.

- CommittableProducerRecords#apply to produce zero or more records within the same transaction as the offset is committed.
- CommittableProducerRecords#one to produce exactly one record within the same transaction as the offset is committed
.

CommittableProducerRecords represents zero or more ProducerRecords and a CommittableOffset, used by TransactionalKafkaProducer to produce the records and commit the offset atomically.

CommittableProducerRecordss can be created using one of the following options.

- CommittableProducerRecords#apply to produce zero or more records within the same transaction as the offset is committed.
- CommittableProducerRecords#one to produce exactly one record within the same transaction as the offset is committed
.

Attributes

Companion
object
Source
CommittableProducerRecords.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
CommittableProducerRecords.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class ConsumerGroupException(groupIds: Set[String]) extends KafkaException

Indicates that one or more of the following conditions occurred while attempting to commit offsets.

- There were CommittableOffsets without a consumer group ID.
- There were CommittableOffsets for multiple consumer group IDs
.

Indicates that one or more of the following conditions occurred while attempting to commit offsets.

- There were CommittableOffsets without a consumer group ID.
- There were CommittableOffsets for multiple consumer group IDs
.

Attributes

Source
ConsumerGroupException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class ConsumerRecord[+K, +V]

ConsumerRecord represents a record which has been consumed from Kafka.

ConsumerRecord represents a record which has been consumed from Kafka. At the very least, this includes a key of type K, value of type V, and the topic, partition, and offset of the consumed record.

To create a new instance, use ConsumerRecord#apply

Attributes

Companion
object
Source
ConsumerRecord.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
ConsumerRecord.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class ConsumerSettings[F[_], K, V]

ConsumerSettings contain settings necessary to create a KafkaConsumer.

ConsumerSettings contain settings necessary to create a KafkaConsumer. At the very least, this includes key and value deserializers.

The following consumer configuration defaults are used.
- auto.offset.reset is set to none to avoid the surprise of the otherwise default latest setting.
- enable.auto.commit is set to false since offset commits are managed manually.

Several convenience functions are provided so that you don't have to work with String values and ConsumerConfig for configuration. It's still possible to specify ConsumerConfig values with functions like withProperty.

ConsumerSettings instances are immutable and all modification functions return a new ConsumerSettings instance.

Use ConsumerSettings#apply to create a new instance.

Attributes

Companion
object
Source
ConsumerSettings.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
ConsumerSettings.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class ConsumerShutdownException extends KafkaException

ConsumerShutdownException indicates that a request could not be completed because the consumer has already shutdown.

ConsumerShutdownException indicates that a request could not be completed because the consumer has already shutdown.

Attributes

Source
ConsumerShutdownException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class DeserializationException(message: String) extends KafkaException

Exception raised with Deserializer#failWith when deserialization was unable to complete successfully.

Exception raised with Deserializer#failWith when deserialization was unable to complete successfully.

Attributes

Source
DeserializationException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class GenericDeserializer[-T <: KeyOrValue, F[_], A]

Functional composable Kafka key- and record deserializer with support for effect types.

Functional composable Kafka key- and record deserializer with support for effect types.

Attributes

Companion
object
Source
Deserializer.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
Deserializer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class GenericSerializer[-T <: KeyOrValue, F[_], A]

Attributes

Companion
object
Source
Serializer.scala
Supertypes
class Object
trait Matchable
class Any

Functional composable Kafka key- and record serializer with support for effect types.

Functional composable Kafka key- and record serializer with support for effect types.

Attributes

Companion
class
Source
Serializer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class Header extends Header

Header represents a String key and Array[Byte] value which can be included as part of Headers when creating a ProducerRecord.

Header represents a String key and Array[Byte] value which can be included as part of Headers when creating a ProducerRecord. Headers are included together with a record once produced, and can be used by consumers.

To create a new Header, use Header#apply.

Attributes

Companion
object
Source
Header.scala
Supertypes
trait Header
class Object
trait Matchable
class Any
object Header

Attributes

Companion
class
Source
Header.scala
Supertypes
class Object
trait Matchable
class Any
Self type
Header.type
sealed abstract class HeaderDeserializer[A]

HeaderDeserializer is a functional deserializer for Kafka record header values.

HeaderDeserializer is a functional deserializer for Kafka record header values. It's similar to Deserializer, except it only has access to the header bytes, and it does not interoperate with the Kafka Deserializer interface.

Attributes

Companion
object
Source
HeaderDeserializer.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
HeaderDeserializer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class HeaderSerializer[A]

HeaderSerializer is a functional serializer for Kafka record header values.

HeaderSerializer is a functional serializer for Kafka record header values. It's similar to Serializer, except it only has access to the value, and it does not interoperate with the Kafka Serializer interface.

Attributes

Companion
object
Source
HeaderSerializer.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
HeaderSerializer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class Headers

Headers represent an immutable append-only collection of Headers.

Headers represent an immutable append-only collection of Headers. To create a new Headers instance, you can use Headers#apply or Headers#empty and add an instance of Header using append.

Attributes

Companion
object
Source
Headers.scala
Supertypes
class Object
trait Matchable
class Any
object Headers

Attributes

Companion
class
Source
Headers.scala
Supertypes
class Object
trait Matchable
class Any
Self type
Headers.type
sealed abstract class IsolationLevel

The available options for ConsumerSettings#withIsolationLevel.

Available options include:
- IsolationLevel#ReadCommitted to only read committed records,
- IsolationLevel#ReadUncommitted to also read uncommitted records
.

The available options for ConsumerSettings#withIsolationLevel.

Available options include:
- IsolationLevel#ReadCommitted to only read committed records,
- IsolationLevel#ReadUncommitted to also read uncommitted records
.

Attributes

Companion
object
Source
IsolationLevel.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
IsolationLevel.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class Jitter[F[_]]

Jitter represents the ability to apply jitter to an existing value n, effectively multiplying n with a pseudorandom value between 0 and 1 (both inclusive, although implementation dependent).

The default Jitter#default uses java.util.Random for pseudorandom values and always applies jitter with a value between 0 (inclusive) and 1 (exclusive)
.

Jitter represents the ability to apply jitter to an existing value n, effectively multiplying n with a pseudorandom value between 0 and 1 (both inclusive, although implementation dependent).

The default Jitter#default uses java.util.Random for pseudorandom values and always applies jitter with a value between 0 (inclusive) and 1 (exclusive)
.
If no jitter is desired, use Jitter#none.

Attributes

Companion
object
Source
Jitter.scala
Supertypes
class Object
trait Matchable
class Any
object Jitter

Attributes

Companion
class
Source
Jitter.scala
Supertypes
class Object
trait Matchable
class Any
Self type
Jitter.type
sealed abstract class KafkaAdminClient[F[_]]

KafkaAdminClient represents an admin client for Kafka, which is able to describe queries about topics, consumer groups, offsets, and other entities related to Kafka.

Use KafkaAdminClient.resource or KafkaAdminClient.stream to create an instance
.

KafkaAdminClient represents an admin client for Kafka, which is able to describe queries about topics, consumer groups, offsets, and other entities related to Kafka.

Use KafkaAdminClient.resource or KafkaAdminClient.stream to create an instance
.

Attributes

Companion
object
Source
KafkaAdminClient.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
KafkaAdminClient.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class KafkaConsumer[F[_], K, V] extends KafkaConsume[F, K, V], KafkaAssignment[F], KafkaOffsetsV2[F], KafkaSubscription[F], KafkaTopics[F], KafkaCommit[F], KafkaMetrics[F], KafkaConsumerLifecycle[F]

KafkaConsumer represents a consumer of Kafka records, with the ability to subscribe to topics, start a single top-level stream, and optionally control it via the provided fiber instance.

The following top-level streams are provided.

- stream provides a single stream of records, where the order of records is guaranteed per topic-partition.
- partitionedStream provides a stream with elements as streams that continually request records for a single partition
.

KafkaConsumer represents a consumer of Kafka records, with the ability to subscribe to topics, start a single top-level stream, and optionally control it via the provided fiber instance.

The following top-level streams are provided.

- stream provides a single stream of records, where the order of records is guaranteed per topic-partition.
- partitionedStream provides a stream with elements as streams that continually request records for a single partition
.
Order is guaranteed per topic-partition, but all assigned partitions will have to be processed in parallel.
- partitionsMapStream provides a stream where each element contains a current assignment. The current assignment is the Map, where keys is a TopicPartition, and values are streams with records for a particular TopicPartition.
For the streams, records are wrapped in CommittableConsumerRecords which provide CommittableOffsets with the ability to commit record offsets to Kafka. For performance reasons, offsets are usually committed in batches using CommittableOffsetBatch. Provided Pipes, like commitBatchWithin are available for batch committing offsets. If you are not committing offsets to Kafka, you can simply discard the CommittableOffset, and only make use of the record.

While it's technically possible to start more than one stream from a single KafkaConsumer, it is generally not recommended as there is no guarantee which stream will receive which records, and there might be an overlap, in terms of duplicate records, between the two streams. If a first stream completes, possibly with error, there's no guarantee the stream has processed all of the records it received, and a second stream from the same KafkaConsumer might not be able to pick up where the first one left off. Therefore, only create a single top-level stream per KafkaConsumer, and if you want to start a new stream if the first one finishes, let the KafkaConsumer shutdown and create a new one.

Attributes

Companion
object
Source
KafkaConsumer.scala
Supertypes
trait KafkaMetrics[F]
trait KafkaCommit[F]
trait KafkaTopics[F]
trait KafkaOffsetsV2[F]
trait KafkaOffsets[F]
trait KafkaAssignment[F]
trait KafkaConsume[F, K, V]
class Object
trait Matchable
class Any
Show all
object KafkaConsumer

Attributes

Companion
class
Source
KafkaConsumer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
abstract class KafkaProducer[F[_], K, V]

KafkaProducer represents a producer of Kafka records, with the ability to produce ProducerRecords using produce.

KafkaProducer represents a producer of Kafka records, with the ability to produce ProducerRecords using produce.

Attributes

Companion
object
Source
KafkaProducer.scala
Supertypes
class Object
trait Matchable
class Any
Known subtypes
class Metrics[F, K, V]
class PartitionsFor[F, K, V]
object KafkaProducer

Attributes

Companion
class
Source
KafkaProducer.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class KafkaProducerConnection[F[_]]

KafkaProducerConnection represents a connection to a Kafka broker that can be used to create KafkaProducer instances.

KafkaProducerConnection represents a connection to a Kafka broker that can be used to create KafkaProducer instances. All KafkaProducer instances created from an given KafkaProducerConnection share a single underlying connection.

Attributes

Companion
object
Source
KafkaProducerConnection.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
KafkaProducerConnection.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed trait Key extends KeyOrValue

Attributes

Source
package.scala
Supertypes
trait KeyOrValue
class Object
trait Matchable
class Any
sealed trait KeyOrValue

Phantom types to indicate whether a Serializer/Deserializer if for keys, values, or both

Phantom types to indicate whether a Serializer/Deserializer if for keys, values, or both

Attributes

Source
package.scala
Supertypes
class Object
trait Matchable
class Any
Known subtypes
trait Key
trait Value
sealed abstract class NotSubscribedException extends KafkaException

NotSubscribedException indicates that a Stream was started in KafkaConsumer even though the consumer had not been subscribed to any topics before starting.

NotSubscribedException indicates that a Stream was started in KafkaConsumer even though the consumer had not been subscribed to any topics before starting.

Attributes

Source
NotSubscribedException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class ProducerRecord[+K, +V]

ProducerRecord represents a record which can be produced to Kafka.

ProducerRecord represents a record which can be produced to Kafka. At the very least, this includes a key of type K, a value of type V, and to which topic the record should be produced. The partition, timestamp, and headers can be set by using the withPartition, withTimestamp, and withHeaders functions, respectively.

To create a new instance, use ProducerRecord#apply.

Attributes

Companion
object
Source
ProducerRecord.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
ProducerRecord.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type

Attributes

Source
package.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class ProducerSettings[F[_], K, V]

ProducerSettings contain settings necessary to create a KafkaProducer.

ProducerSettings contain settings necessary to create a KafkaProducer. At the very least, this includes a key serializer and a value serializer.

Several convenience functions are provided so that you don't have to work with String values and ProducerConfig for configuration. It's still possible to specify ProducerConfig values with functions like withProperty.

ProducerSettings instances are immutable and all modification functions return a new ProducerSettings instance.

Use ProducerSettings#apply to create a new instance.

Attributes

Companion
object
Source
ProducerSettings.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
ProducerSettings.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class SerializationException(message: String) extends KafkaException

Exception raised with Serializer#failWith when serialization was unable to complete successfully.

Exception raised with Serializer#failWith when serialization was unable to complete successfully.

Attributes

Source
SerializationException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed abstract class Timestamp

Timestamp is an optional timestamp value representing a createTime, logAppendTime, unknownTime, or no timestamp at all.

Timestamp is an optional timestamp value representing a createTime, logAppendTime, unknownTime, or no timestamp at all.

Attributes

Companion
object
Source
Timestamp.scala
Supertypes
class Object
trait Matchable
class Any
object Timestamp

Attributes

Companion
class
Source
Timestamp.scala
Supertypes
class Object
trait Matchable
class Any
Self type
Timestamp.type
abstract class TransactionalKafkaProducer[F[_], K, V]

Represents a producer of Kafka records specialized for 'read-process-write' streams, with the ability to atomically produce ProducerRecords and commit corresponding CommittableOffsets using produce.

Records are wrapped in TransactionalProducerRecords, which is a chunk of CommittableProducerRecord which wrap zero or more records together with a CommittableOffset
.

Represents a producer of Kafka records specialized for 'read-process-write' streams, with the ability to atomically produce ProducerRecords and commit corresponding CommittableOffsets using produce.

Records are wrapped in TransactionalProducerRecords, which is a chunk of CommittableProducerRecord which wrap zero or more records together with a CommittableOffset
.

Attributes

Companion
object
Source
TransactionalKafkaProducer.scala
Supertypes
class Object
trait Matchable
class Any
Known subtypes
class Metrics[F, K, V]
class WithoutOffsets[F, K, V]

Attributes

Companion
class
Source
TransactionalKafkaProducer.scala
Supertypes
class Object
trait Matchable
class Any
Self type

Attributes

Source
package.scala
Supertypes
class Object
trait Matchable
class Any
Self type
sealed abstract class TransactionalProducerSettings[F[_], K, V]

TransactionalProducerSettings contain settings necessary to create a TransactionalKafkaProducer.

TransactionalProducerSettings contain settings necessary to create a TransactionalKafkaProducer. This includes a transactional ID and any other ProducerSettings.

TransactionalProducerSettings instances are immutable and modification functions return a new TransactionalProducerSettings instance.

Use TransactionalProducerSettings.apply to create a new instance.

Attributes

Companion
object
Source
TransactionalProducerSettings.scala
Supertypes
class Object
trait Matchable
class Any

Attributes

Companion
class
Source
TransactionalProducerSettings.scala
Supertypes
trait Sum
trait Mirror
class Object
trait Matchable
class Any
Self type
sealed abstract class UnexpectedTopicException(topic: String) extends KafkaException

UnexpectedTopicException is raised when serialization or deserialization occurred for an unexpected topic which isn't supported by the Serializer or Deserializer.

UnexpectedTopicException is raised when serialization or deserialization occurred for an unexpected topic which isn't supported by the Serializer or Deserializer.

Attributes

Source
UnexpectedTopicException.scala
Supertypes
class KafkaException
class Exception
class Throwable
trait Serializable
class Object
trait Matchable
class Any
Show all
sealed trait Value extends KeyOrValue

Attributes

Source
package.scala
Supertypes
trait KeyOrValue
class Object
trait Matchable
class Any
object instances

Attributes

Source
instances.scala
Supertypes
class Object
trait Matchable
class Any
Self type
instances.type

Types

Attributes

Source
package.scala
type Id[+A] = A

Attributes

Source
package.scala
type KafkaByteConsumer = Consumer[Array[Byte], Array[Byte]]

Alias for Java Kafka Consumer[Array[Byte], Array[Byte]].

Alias for Java Kafka Consumer[Array[Byte], Array[Byte]].

Attributes

Source
package.scala
type KafkaByteConsumerRecord = ConsumerRecord[Array[Byte], Array[Byte]]

Alias for Java Kafka ConsumerRecord[Array[Byte], Array[Byte]].

Alias for Java Kafka ConsumerRecord[Array[Byte], Array[Byte]].

Attributes

Source
package.scala
type KafkaByteConsumerRecords = ConsumerRecords[Array[Byte], Array[Byte]]

Alias for Java Kafka ConsumerRecords[Array[Byte], Array[Byte]].

Alias for Java Kafka ConsumerRecords[Array[Byte], Array[Byte]].

Attributes

Source
package.scala
type KafkaByteProducer = Producer[Array[Byte], Array[Byte]]

Alias for Java Kafka Producer[Array[Byte], Array[Byte]].

Alias for Java Kafka Producer[Array[Byte], Array[Byte]].

Attributes

Source
package.scala
type KafkaByteProducerRecord = ProducerRecord[Array[Byte], Array[Byte]]

Alias for Java Kafka ProducerRecord[Array[Byte], Array[Byte]].

Alias for Java Kafka ProducerRecord[Array[Byte], Array[Byte]].

Attributes

Source
package.scala
type KafkaDeserializer[A] = Deserializer[A]

Alias for Java Kafka Deserializer[A].

Alias for Java Kafka Deserializer[A].

Attributes

Source
package.scala
type KafkaHeader = Header

Alias for Java Kafka Header.

Alias for Java Kafka Header.

Attributes

Source
package.scala
type KafkaHeaders = Headers

Alias for Java Kafka Headers.

Alias for Java Kafka Headers.

Attributes

Source
package.scala
type KafkaSerializer[A] = Serializer[A]

Alias for Java Kafka Serializer[A].

Alias for Java Kafka Serializer[A].

Attributes

Source
package.scala
type KeyDeserializer[F[_], A] = GenericDeserializer[Key, F, A]

Attributes

Source
package.scala
type KeySerializer[F[_], A] = GenericSerializer[Key, F, A]

Attributes

Source
package.scala
type ProducerRecords[K, V] = Chunk[ProducerRecord[K, V]]

Attributes

Source
package.scala
type ProducerResult[K, V] = Chunk[(ProducerRecord[K, V], RecordMetadata)]

Attributes

Source
package.scala
type Serializer[F[_], A] = GenericSerializer[KeyOrValue, F, A]

Attributes

Source
package.scala

Attributes

Source
package.scala

Attributes

Source
package.scala
type ValueSerializer[F[_], A] = GenericSerializer[Value, F, A]

Attributes

Source
package.scala

Value members

Concrete methods

def commitBatchWithin[F[_]](n: Int, d: FiniteDuration)(implicit F: Temporal[F]): (F, CommittableOffset[F]) => Unit

Commits offsets in batches of every n offsets or time window of length d, whichever happens first.

Commits offsets in batches of every n offsets or time window of length d, whichever happens first. If there are no offsets to commit within a time window, no attempt will be made to commit offsets for that time window.

Attributes

Source
package.scala

Concrete fields

Attributes

Source
package.scala

Attributes

Source
package.scala