- AbstractProcessor<K,V> - Class in org.apache.kafka.streams.processor
-
- AbstractProcessor() - Constructor for class org.apache.kafka.streams.processor.AbstractProcessor
-
- accepts(StateStore) - Method in interface org.apache.kafka.streams.state.QueryableStoreType
-
Called when searching for
StateStore
s to see if they
match the type expected by implementors of this interface
- addGlobalStore(StateStoreSupplier<KeyValueStore>, String, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- addGlobalStore(StateStoreSupplier<KeyValueStore>, String, TimestampExtractor, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- addInternalTopic(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds an internal topic
- addLatencyAndThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Add a latency and throughput sensor for a specific operation, which will include the following sensors:
average latency
max latency
throughput (num.operations / time unit)
Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the
same scope if it has not been created.
- addProcessor(String, ProcessorSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new processor node that receives and processes records output by one or more parent source or processor node.
- addSensor(String, Sensor.RecordingLevel) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Generic method to create a sensor.
- addSensor(String, Sensor.RecordingLevel, Sensor...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Generic method to create a sensor with parent sensors.
- addSink(String, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, StreamPartitioner, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic, using
the supplied partitioner.
- addSink(String, String, Serializer, Serializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, Serializer<K>, Serializer<V>, StreamPartitioner<? super K, ? super V>, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSource(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addSource(String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forwards the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forwards the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forwards the records to child processor and/or sink nodes.
- addStateStore(StateStoreSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds a state store
- addThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Add a throughput sensor for a specific operation:
throughput (num.operations / time unit)
Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the
same scope if it has not been created.
- advanceBy(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return a window definition with the original size, but advance ("hop") the window by the given interval, which
specifies by how much a window moves forward relative to the previous one.
- advanceMs - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the window's advance interval in milliseconds, i.e., by how much a window moves forward relative to
the previous one.
- after(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Changes the end window boundary to timeDifferenceMs
but keep the start window boundary as is.
- afterMs - Variable in class org.apache.kafka.streams.kstream.JoinWindows
-
Maximum time difference for tuples that are after the join tuple.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined windows.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined windows.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined windows.
- aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined
SessionWindows
.
- aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined
SessionWindows
.
- aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined
SessionWindows
.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- Aggregator<K,V,VA> - Interface in org.apache.kafka.streams.kstream
-
The Aggregator
interface for aggregating values of the given key.
- all() - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Return an iterator over all keys in this store.
- allMetadata() - Method in class org.apache.kafka.streams.KafkaStreams
-
Find all currently running
KafkaStreams
instances (potentially remotely) that use the same
application ID
as this instance (i.e., all instances that belong to
the same Kafka Streams application) and return
StreamsMetadata
for each discovered instance.
- allMetadataForStore(String) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find all currently running
KafkaStreams
instances (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and return
StreamsMetadata
for each discovered instance.
- appConfigs() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns all the application config properties as key/value pairs.
- appConfigsWithPrefix(String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns all the application config properties with the given key prefix, as key/value pairs
stripping the prefix.
- APPLICATION_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
application.id
- APPLICATION_SERVER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
user.endpoint
- applicationId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the application id
- apply(K, V, VA) - Method in interface org.apache.kafka.streams.kstream.Aggregator
-
Compute a new aggregate from the key and value of a record and the current aggregate of the same key.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.ForeachAction
-
Perform an action for each record of a stream.
- apply() - Method in interface org.apache.kafka.streams.kstream.Initializer
-
Return the initial value for an aggregation.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.KeyValueMapper
-
Map a record with the given key and value to a new value.
- apply(K, V, V) - Method in interface org.apache.kafka.streams.kstream.Merger
-
Compute a new aggregate from the key and two aggregates.
- apply(K, V) - Method in class org.apache.kafka.streams.kstream.PrintForeachAction
-
- apply(V, V) - Method in interface org.apache.kafka.streams.kstream.Reducer
-
Aggregate the two given values into a single one.
- apply(V1, V2) - Method in interface org.apache.kafka.streams.kstream.ValueJoiner
-
Return a joined value consisting of value1
and value2
.
- apply(V) - Method in interface org.apache.kafka.streams.kstream.ValueMapper
-
Map the given value to a new value.
- approximateNumEntries() - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Return an approximate count of key-value mappings in this store.
- AT_LEAST_ONCE - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- CACHE_MAX_BYTES_BUFFERING_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
cache.max.bytes.buffering
- cleanUp() - Method in class org.apache.kafka.streams.KafkaStreams
-
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
client.id
- close() - Method in class org.apache.kafka.streams.KafkaStreams
-
Shutdown this KafkaStreams
instance by signaling all the threads to stop, and then wait for them to join.
- close(long, TimeUnit) - Method in class org.apache.kafka.streams.KafkaStreams
-
Shutdown this KafkaStreams
by signaling all the threads to stop, and then wait up to the timeout for the
threads to join.
- close() - Method in class org.apache.kafka.streams.kstream.PrintForeachAction
-
- close() - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Close this processor and clean up any resources.
- close() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.Processor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Close the storage engine.
- close() - Method in interface org.apache.kafka.streams.state.KeyValueIterator
-
- close() - Method in interface org.apache.kafka.streams.state.WindowStoreIterator
-
- commit() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Requests a commit
- COMMIT_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
commit.interval.ms
- compareTo(TaskId) - Method in class org.apache.kafka.streams.processor.TaskId
-
- configDef() - Static method in class org.apache.kafka.streams.StreamsConfig
-
Return a copy of the config definition.
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
connections.max.idle.ms
- connectProcessorAndStateStores(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects the processor and the state stores
- connectProcessors(String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects a list of processors.
- connectSourceStoreAndTopic(String, String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
This is used only for KStreamBuilder: when adding a KTable from a source topic,
we need to add the topic as the KTable's materialized state store's changelog.
- CONSUMER_PREFIX - Static variable in class org.apache.kafka.streams.StreamsConfig
-
Prefix used to isolate consumer
configs from producer
configs.
- consumerPrefix(String) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- context() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- copartitionGroups() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the copartition groups.
- copartitionSources(Collection<String>) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Asserts that the streams of the specified source nodes must be copartitioned.
- count(String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key.
- count() - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key.
- count(StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key.
- count(Windows<W>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key and the defined windows.
- count(Windows<W>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key and the defined windows.
- count(Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key and the defined windows.
- count(SessionWindows, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key into
SessionWindows
.
- count(SessionWindows) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key into
SessionWindows
.
- count(SessionWindows, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key into
SessionWindows
.
- count(String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of the original
KTable
that got
mapped
to
the same key into a new instance of
KTable
.
- count() - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of the original
KTable
that got
mapped
to
the same key into a new instance of
KTable
.
- count(StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of the original
KTable
that got
mapped
to
the same key into a new instance of
KTable
.
- create(StateStoreProvider, String) - Method in interface org.apache.kafka.streams.state.QueryableStoreType
-
Create an instance of T (usually a facade) that developers can use
to query the underlying
StateStore
s
- create(String) - Static method in class org.apache.kafka.streams.state.Stores
-
- FailOnInvalidTimestamp - Class in org.apache.kafka.streams.processor
-
Retrieves embedded metadata timestamps from Kafka messages.
- FailOnInvalidTimestamp() - Constructor for class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
-
- fetch(K) - Method in interface org.apache.kafka.streams.state.ReadOnlySessionStore
-
Retrieve all aggregated sessions for the provided key
For each key, the iterator guarantees ordering of sessions, starting from the oldest/earliest
available session to the newest/latest session.
- fetch(K, K) - Method in interface org.apache.kafka.streams.state.ReadOnlySessionStore
-
Retrieve all aggregated sessions for the given range of keys
For each key, the iterator guarantees ordering of sessions, starting from the oldest/earliest
available session to the newest/latest session.
- fetch(K, long, long) - Method in interface org.apache.kafka.streams.state.ReadOnlyWindowStore
-
Get all the key-value pairs with the given key and the time range from all
the existing windows.
- fetch(K, K, long, long) - Method in interface org.apache.kafka.streams.state.ReadOnlyWindowStore
-
Get all the key-value pairs in the given key range and time range from all
the existing windows.
- filter(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
that consists of all records of this stream which satisfy the given predicate.
- filter(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists of all records of this KTable
which satisfy the given
predicate.
- filter(Predicate<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists of all records of this KTable
which satisfy the given
predicate.
- filter(Predicate<? super K, ? super V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists of all records of this KTable
which satisfy the given
predicate.
- filterNot(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
that consists all records of this stream which do not satisfy the given
predicate.
- filterNot(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists all records of this KTable
which do not satisfy the
given predicate.
- filterNot(Predicate<? super K, ? super V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists all records of this KTable
which do not satisfy the
given predicate.
- filterNot(Predicate<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists all records of this KTable
which do not satisfy the
given predicate.
- findSessions(K, long, long) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Fetch any sessions with the matching key and the sessions end is ≥ earliestSessionEndTime and the sessions
start is ≤ latestSessionStartTime
- findSessions(K, K, long, long) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Fetch any sessions in the given range of keys and the sessions end is ≥ earliestSessionEndTime and the sessions
start is ≤ latestSessionStartTime
- flatMap(KeyValueMapper<? super K, ? super V, ? extends Iterable<? extends KeyValue<? extends KR, ? extends VR>>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into zero or more records in the output stream (both key and value type
can be altered arbitrarily).
- flatMapValues(ValueMapper<? super V, ? extends Iterable<? extends VR>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
by transforming the value of each record in this stream into zero or more values
with the same key in the new stream.
- flush() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Flush any cached data
- foreach(ForeachAction<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Perform an action on each record of KStream
.
- foreach(ForeachAction<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- ForeachAction<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
ForeachAction
interface for performing an action on a
key-value
pair
.
- forward(K, V) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to the downstream processors
- forward(K, V, int) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by childIndex
- forward(K, V, String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by the downstream processor name
- KafkaClientSupplier - Interface in org.apache.kafka.streams
-
KafkaClientSupplier
can be used to provide custom Kafka clients to a
KafkaStreams
instance.
- KafkaStreams - Class in org.apache.kafka.streams
-
A Kafka client that allows for performing continuous computation on input coming from one or more input topics and
sends output to zero, one, or more output topics.
- KafkaStreams(TopologyBuilder, Properties) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams(TopologyBuilder, StreamsConfig) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams(TopologyBuilder, StreamsConfig, KafkaClientSupplier) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams.State - Enum in org.apache.kafka.streams
-
Kafka Streams states are the possible state that a Kafka Streams instance can be in.
- KafkaStreams.StateListener - Interface in org.apache.kafka.streams
-
- key - Variable in class org.apache.kafka.streams.KeyValue
-
The key of the key-value pair.
- key() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the key of the window.
- KEY_SERDE_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- keyDeserializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key deserializer.
- keyFrom(byte[]) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Deserialize the key from raw bytes.
- keySerde() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the default key serde
- keySerde() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serde.
- keySerde() - Method in class org.apache.kafka.streams.StreamsConfig
-
Deprecated.
- keySerializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serializer.
- KeyValue<K,V> - Class in org.apache.kafka.streams
-
A key-value pair defined for a single Kafka Streams record.
- KeyValue(K, V) - Constructor for class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- KeyValueIterator<K,V> - Interface in org.apache.kafka.streams.state
-
- KeyValueMapper<K,V,VR> - Interface in org.apache.kafka.streams.kstream
-
The
KeyValueMapper
interface for mapping a
key-value pair
to a new value of arbitrary type.
- KeyValueStore<K,V> - Interface in org.apache.kafka.streams.state
-
A key-value store that supports put/get/delete and range queries.
- keyValueStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- KGroupedStream<K,V> - Interface in org.apache.kafka.streams.kstream
-
KGroupedStream
is an abstraction of a
grouped record stream of
KeyValue
pairs.
- KGroupedTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KGroupedTable
is an abstraction of a re-grouped changelog stream from a primary-keyed table,
usually on a different grouping key than the original primary key.
- KStream<K,V> - Interface in org.apache.kafka.streams.kstream
-
KStream
is an abstraction of a
record stream of
KeyValue
pairs, i.e., each record is an
independent entity/event in the real world.
- KStreamBuilder - Class in org.apache.kafka.streams.kstream
-
KStreamBuilder
provide the high-level Kafka Streams DSL to specify a Kafka Streams topology.
- KStreamBuilder() - Constructor for class org.apache.kafka.streams.kstream.KStreamBuilder
-
- KTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KTable
is an abstraction of a changelog stream from a primary-keyed table.
- main(String[]) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- maintainMs() - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.SessionWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.Windows
-
Return the window maintain duration (retention time) in milliseconds.
- map(KeyValueMapper<? super K, ? super V, ? extends KeyValue<? extends KR, ? extends VR>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into a new record in the output stream (both key and value type can be
altered arbitrarily).
- mapValues(ValueMapper<? super V, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform the value of each input record into a new value (with possible new type) of the output record.
- mapValues(ValueMapper<? super V, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
by transforming the value of each record in this KTable
into a new value
(with possible new type)in the new KTable
.
- mapValues(ValueMapper<? super V, ? extends VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
by transforming the value of each record in this KTable
into a new value
(with possible new type)in the new KTable
.
- mapValues(ValueMapper<? super V, ? extends VR>, Serde<VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
by transforming the value of each record in this KTable
into a new value
(with possible new type)in the new KTable
.
- maxEntries(int) - Method in interface org.apache.kafka.streams.state.Stores.InMemoryKeyValueFactory
-
Limits the in-memory key-value store to hold a maximum number of entries.
- maxNumPartitions(Cluster, Set<String>) - Method in class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
- merge(KStream<K, V>...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
- Merger<K,V> - Interface in org.apache.kafka.streams.kstream
-
The interface for merging aggregate values for
SessionWindows
with the given key.
- METADATA_MAX_AGE_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metadata.max.age.ms
- metadataForKey(String, K, Serializer<K>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find the currently running
KafkaStreams
instance (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and the
StateStore
contains the given
key
and return
StreamsMetadata
for it.
- metadataForKey(String, K, StreamPartitioner<? super K, ?>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find the currently running
KafkaStreams
instance (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and the
StateStore
contains the given
key
and return
StreamsMetadata
for it.
- METRIC_REPORTER_CLASSES_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metric.reporters
- metrics() - Method in class org.apache.kafka.streams.KafkaStreams
-
Get read-only handle on global metrics registry.
- metrics() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns Metrics instance
- metrics() - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Get read-only handle on global metrics registry.
- METRICS_NUM_SAMPLES_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.num.samples
- METRICS_RECORDING_LEVEL_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.record.level
- METRICS_SAMPLE_WINDOW_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.sample.window.ms
- of(long) - Static method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are within timeDifferenceMs
,
i.e., the timestamp of a record from the secondary stream is max timeDifferenceMs
earlier or later than
the timestamp of the record from the primary stream.
- of(long) - Static method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return a window definition with the given window size, and with the advance interval being equal to the window
size.
- of() - Static method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return an unlimited window starting at timestamp zero.
- offset() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the offset of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- onChange(KafkaStreams.State, KafkaStreams.State) - Method in interface org.apache.kafka.streams.KafkaStreams.StateListener
-
Called when state changes.
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
-
Raises an exception on every call.
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.LogAndSkipOnInvalidTimestamp
-
Writes a log WARN message when the extracted timestamp is invalid (negative) but returns the invalid timestamp as-is,
which ultimately causes the record to be skipped and not to be processed.
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.UsePreviousTimeOnInvalidTimestamp
-
Returns the current stream-time as new timestamp for the record.
- org.apache.kafka.streams - package org.apache.kafka.streams
-
- org.apache.kafka.streams.errors - package org.apache.kafka.streams.errors
-
- org.apache.kafka.streams.kstream - package org.apache.kafka.streams.kstream
-
- org.apache.kafka.streams.processor - package org.apache.kafka.streams.processor
-
- org.apache.kafka.streams.state - package org.apache.kafka.streams.state
-
- outerJoin(KStream<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, JoinWindows) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Join records of this stream with another KStream
's records using windowed left equi join with default
serializers and deserializers.
- outerJoin(KStream<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, JoinWindows, Serde<K>, Serde<V>, Serde<VO>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Join records of this stream with another KStream
's records using windowed left equi join.
- outerJoin(KTable<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Join records of this KTable
(left input) with another KTable
's (right input) records using
non-windowed outer equi join.
- outerJoin(KTable<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Join records of this KTable
(left input) with another KTable
's (right input) records using
non-windowed outer equi join.
- outerJoin(KTable<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Join records of this KTable
(left input) with another KTable
's (right input) records using
non-windowed outer equi join.
- overlap(Window) - Method in class org.apache.kafka.streams.kstream.Window
-
Check if the given window overlaps with this window.
- pair(K, V) - Static method in class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- parse(String) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- partition() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the partition id of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- partition(K, V, int) - Method in interface org.apache.kafka.streams.processor.StreamPartitioner
-
Determine the partition number for a record with the given key and value and the current number of partitions.
- partition - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the partition.
- PARTITION_GROUPER_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
partition.grouper
- PartitionGrouper - Interface in org.apache.kafka.streams.processor
-
A partition grouper that generates partition groups given the list of topic-partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
Generate tasks with the assigned topic partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in interface org.apache.kafka.streams.processor.PartitionGrouper
-
Returns a map of task ids to groups of partitions.
- peek(ForeachAction<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Perform an action on each record of KStream
.
- peekNextKey() - Method in interface org.apache.kafka.streams.state.KeyValueIterator
-
Peek at the next key without advancing the iterator
- persistent() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Return if the storage is persistent or not.
- persistent() - Method in interface org.apache.kafka.streams.state.Stores.KeyValueFactory
-
Keep all key-value entries off-heap in a local database, although for durability all entries are recorded in a Kafka
topic that can be read to restore the entries if they are lost.
- POLL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
poll.ms
- port() - Method in class org.apache.kafka.streams.state.HostInfo
-
- port() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- postProcessParsedConfig(Map<String, Object>) - Method in class org.apache.kafka.streams.StreamsConfig
-
- Predicate<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
Predicate
interface represents a predicate (boolean-valued function) of a
KeyValue
pair.
- print() - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print() - Method in interface org.apache.kafka.streams.kstream.KTable
-
- print(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- print(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- PrintForeachAction<K,V> - Class in org.apache.kafka.streams.kstream
-
- PrintForeachAction(PrintWriter, String) - Constructor for class org.apache.kafka.streams.kstream.PrintForeachAction
-
Print data message with given writer.
- process(ProcessorSupplier<? super K, ? super V>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
- process(K, V) - Method in interface org.apache.kafka.streams.processor.Processor
-
Process the record with the given key and value.
- PROCESSING_GUARANTEE_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
cache.max.bytes.buffering
- Processor<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor of key-value pair records.
- ProcessorContext - Interface in org.apache.kafka.streams.processor
-
Processor context interface.
- ProcessorStateException - Exception in org.apache.kafka.streams.errors
-
Indicates a processor state operation (e.g.
- ProcessorStateException(String) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorSupplier<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor supplier that can create one or more
Processor
instances.
- PRODUCER_PREFIX - Static variable in class org.apache.kafka.streams.StreamsConfig
-
Prefix used to isolate producer
configs from consumer
configs.
- producerPrefix(String) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
- punctuate(long) - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- punctuate(long) - Method in interface org.apache.kafka.streams.processor.Processor
-
- put(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key
- put(Windowed<K>, AGG) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Write the aggregated value for the provided key to the store
- put(K, V) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the current wall-clock time as the timestamp
into the corresponding window
- put(K, V, long) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the given timestamp into the corresponding window
- putAll(List<KeyValue<K, V>>) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update all the given key/value pairs
- putIfAbsent(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key, unless a value
is already associated with the key
- range(K, K) - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Get an iterator over a given range of keys.
- rawKey(K) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given key.
- rawValue(V) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given value.
- readFrom(DataInputStream) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- readFrom(ByteBuffer) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- ReadOnlyKeyValueStore<K,V> - Interface in org.apache.kafka.streams.state
-
A key value store that only supports read operations.
- ReadOnlySessionStore<K,AGG> - Interface in org.apache.kafka.streams.state
-
A session store that only supports read operations.
- ReadOnlyWindowStore<K,V> - Interface in org.apache.kafka.streams.state
-
A window store that only supports read operations
Implementations should be thread-safe as concurrent reads and writes
are expected.
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
receive.buffer.bytes
- RECONNECT_BACKOFF_MAX_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
reconnect.backoff.max
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
reconnect.backoff.ms
- recordLatency(Sensor, long, long) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Record the given latency value of the sensor.
- recordThroughput(Sensor, long) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Record the throughput value of a sensor.
- reduce(Reducer<V>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the values of records in this stream by the grouped key.
- reduce(Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the values of records in this stream by the grouped key.
- reduce(Reducer<V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the value of records in this stream by the grouped key.
- reduce(Reducer<V>, Windows<W>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the number of records in this stream by the grouped key and the defined windows.
- reduce(Reducer<V>, Windows<W>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the number of records in this stream by the grouped key and the defined windows.
- reduce(Reducer<V>, Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the values of records in this stream by the grouped key and the defined windows.
- reduce(Reducer<V>, SessionWindows, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
- reduce(Reducer<V>, SessionWindows) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
- reduce(Reducer<V>, SessionWindows, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
- reduce(Reducer<V>, Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
.
- reduce(Reducer<V>, Reducer<V>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
.
- reduce(Reducer<V>, Reducer<V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
.
- Reducer<V> - Interface in org.apache.kafka.streams.kstream
-
The Reducer
interface for combining two values of the same type into a new value.
- register(StateStore, boolean, StateRestoreCallback) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Registers and possibly restores the specified storage engine.
- remove(Windowed<K>) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Remove the session aggregated with provided
Windowed
key from the store
- removeSensor(Sensor) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Remove a sensor.
- repartitionSourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- REPLICATION_FACTOR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
replication.factor
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
request.timeout.ms
- restore(byte[], byte[]) - Method in interface org.apache.kafka.streams.processor.StateRestoreCallback
-
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
retry.backoff.ms
- ROCKSDB_CONFIG_SETTER_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
rocksdb.config.setter
- RocksDBConfigSetter - Interface in org.apache.kafka.streams.state
-
An interface to that allows developers to customize the RocksDB settings
for a given Store.
- schedule(long) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Schedules a periodic operation for processors.
- SECURITY_PROTOCOL_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
security.protocol
- segments - Variable in class org.apache.kafka.streams.kstream.Windows
-
- segments(int) - Method in class org.apache.kafka.streams.kstream.Windows
-
Set the number of segments to be used for rolling the window store.
- selectKey(KeyValueMapper<? super K, ? super V, ? extends KR>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Set a new key (with possibly new type) for each input record.
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
send.buffer.bytes
- sessionStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- SessionStore<K,AGG> - Interface in org.apache.kafka.streams.state
-
Interface for storing the aggregated values of sessions
- sessionWindowed(long) - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
- SessionWindows - Class in org.apache.kafka.streams.kstream
-
A session based window specification used for aggregating events into sessions.
- setApplicationId(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Set the applicationId to be used for auto-generated internal topics.
- setConfig(String, Options, Map<String, Object>) - Method in interface org.apache.kafka.streams.state.RocksDBConfigSetter
-
Set the rocks db options for the provided storeName.
- setStateListener(KafkaStreams.StateListener) - Method in class org.apache.kafka.streams.KafkaStreams
-
- setUncaughtExceptionHandler(Thread.UncaughtExceptionHandler) - Method in class org.apache.kafka.streams.KafkaStreams
-
Set the handler invoked when a
internal thread
abruptly
terminates due to an uncaught exception.
- sinkTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- size() - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
- size() - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
- size() - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return the size of the specified windows in milliseconds.
- size() - Method in class org.apache.kafka.streams.kstream.Windows
-
Return the size of the specified windows in milliseconds.
- sizeMs - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the windows in milliseconds.
- sourceTopicPattern() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- sourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- start() - Method in class org.apache.kafka.streams.KafkaStreams
-
Start the KafkaStreams
instance by starting all its threads.
- start() - Method in class org.apache.kafka.streams.kstream.Window
-
Return the start timestamp of this window.
- startMs - Variable in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
The start timestamp of the window.
- startMs - Variable in class org.apache.kafka.streams.kstream.Window
-
- startOn(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return a new unlimited window for the specified start timestamp.
- state() - Method in class org.apache.kafka.streams.KafkaStreams
-
- STATE_CLEANUP_DELAY_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.cleanup.delay
- STATE_DIR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.dir
- stateChangelogTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- stateDir() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the state directory for the partition.
- StateRestoreCallback - Interface in org.apache.kafka.streams.processor
-
Restoration logic for log-backed state stores upon restart,
it takes one record at a time from the logs to apply to the restoring state.
- StateSerdes<K,V> - Class in org.apache.kafka.streams.state
-
Factory for creating serializers / deserializers for state stores in Kafka Streams.
- StateSerdes(String, Serde<K>, Serde<V>) - Constructor for class org.apache.kafka.streams.state.StateSerdes
-
Create a context for serialization using the specified serializers and deserializers which
must match the key and value types used as parameters for this object; the state changelog topic
is provided to bind this serde factory to, so that future calls for serialize / deserialize do not
need to provide the topic name any more.
- StateStore - Interface in org.apache.kafka.streams.processor
-
A storage engine for managing state maintained by a stream processor.
- stateStoreNames() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- stateStoreNameToSourceTopics() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- StateStoreSupplier<T extends StateStore> - Interface in org.apache.kafka.streams.processor
-
A state store supplier which can create one or more
StateStore
instances.
- store(String, QueryableStoreType<T>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Get a facade wrapping the local
StateStore
instances with the provided
storeName
if the Store's
type is accepted by the provided
queryableStoreType
.
- Stores - Class in org.apache.kafka.streams.state
-
Factory for creating state stores in Kafka Streams.
- Stores() - Constructor for class org.apache.kafka.streams.state.Stores
-
- Stores.InMemoryKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create in-memory key-value stores.
- Stores.KeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
- Stores.PersistentKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create off-heap key-value stores that use a local database.
- Stores.StoreFactory - Class in org.apache.kafka.streams.state
-
- Stores.StoreFactory() - Constructor for class org.apache.kafka.streams.state.Stores.StoreFactory
-
- Stores.ValueFactory<K> - Class in org.apache.kafka.streams.state
-
The factory for creating off-heap key-value stores.
- Stores.ValueFactory() - Constructor for class org.apache.kafka.streams.state.Stores.ValueFactory
-
- stream(String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TopologyBuilder.AutoOffsetReset, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TopologyBuilder.AutoOffsetReset, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TimestampExtractor, Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TopologyBuilder.AutoOffsetReset, TimestampExtractor, Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TimestampExtractor, Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TopologyBuilder.AutoOffsetReset, TimestampExtractor, Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- STREAM_THREAD_INSTANCE - Static variable in class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamPartitioner<K,V> - Interface in org.apache.kafka.streams.processor
-
Determine how records are distributed among the partitions in a Kafka topic.
- StreamsConfig - Class in org.apache.kafka.streams
-
- StreamsConfig(Map<?, ?>) - Constructor for class org.apache.kafka.streams.StreamsConfig
-
Create a new StreamsConfig
using the given properties.
- StreamsConfig.InternalConfig - Class in org.apache.kafka.streams
-
- StreamsConfig.InternalConfig() - Constructor for class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamsException - Exception in org.apache.kafka.streams.errors
-
- StreamsException(String) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsMetadata - Class in org.apache.kafka.streams.state
-
Represents the state of an instance (process) in a
KafkaStreams
application.
- StreamsMetadata(HostInfo, Set<String>, Set<TopicPartition>) - Constructor for class org.apache.kafka.streams.state.StreamsMetadata
-
- StreamsMetrics - Interface in org.apache.kafka.streams
-
The Kafka Streams metrics interface for adding metric sensors and collecting metric values.
- subscriptionUpdates() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- table(String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(String, StateStoreSupplier<KeyValueStore>) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, String, StateStoreSupplier<KeyValueStore>) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TimestampExtractor, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, TimestampExtractor, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(Serde<K>, Serde<V>, String, StateStoreSupplier<KeyValueStore>) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(Serde<K>, Serde<V>, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TimestampExtractor, Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, TimestampExtractor, Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, TimestampExtractor, Serde<K>, Serde<V>, String, StateStoreSupplier<KeyValueStore>) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- TaskAssignmentException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying to assign
stream tasks
to
threads
.
- TaskAssignmentException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- taskId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the task id
- TaskId - Class in org.apache.kafka.streams.processor
-
The task ID representation composed as topic group ID plus the assigned partition ID.
- TaskId(int, int) - Constructor for class org.apache.kafka.streams.processor.TaskId
-
- TaskIdFormatException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying parse the
task id
from the read string.
- TaskIdFormatException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- TaskIdFormatException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- TaskIdFormatException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- test(K, V) - Method in interface org.apache.kafka.streams.kstream.Predicate
-
Test if the record with the given key and value satisfies the predicate.
- through(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new KStream
from the topic using default serializers and
deserializers and producer's DefaultPartitioner
.
- through(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new
KStream
from the topic using default serializers and
deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, and creates a new KStream
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new
KStream
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic using default
serializers and deserializers and producer's DefaultPartitioner
.
- through(String, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic using default
serializers and deserializers and producer's DefaultPartitioner
.
- through(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic using default
serializers and deserializers and producer's DefaultPartitioner
.
- through(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using default
serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of
records to partitions.
- through(StreamPartitioner<? super K, ? super V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using default
serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of
records to partitions.
- through(StreamPartitioner<? super K, ? super V>, String, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using default
serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of
records to partitions.
- through(Serde<K>, Serde<V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic.
- through(Serde<K>, Serde<V>, String, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic.
- through(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- timestamp() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the current timestamp.
- TIMESTAMP_EXTRACTOR_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- TimestampExtractor - Interface in org.apache.kafka.streams.processor
-
An interface that allows the Kafka Streams framework to extract a timestamp from an instance of ConsumerRecord
.
- TimeWindows - Class in org.apache.kafka.streams.kstream
-
The fixed-size time-based window specifications used for aggregations.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config and producer's
DefaultPartitioner
.
- to(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using a customizable
StreamPartitioner
to determine the distribution
of records to partitions.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using default serializers and deserializers and producer's
DefaultPartitioner
.
- to(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using default serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using a customizable
StreamPartitioner
to determine the
distribution of records to partitions.
- topic() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the topic name of the current input record; could be null if it is not
available (for example, if this method is invoked from the punctuate call)
- topic() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the topic.
- topicGroupId - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the topic group.
- topicGroups() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the map of topic groups keyed by the group id.
- topicPartitions() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- TopologyBuilder - Class in org.apache.kafka.streams.processor
-
A component that is used to build a ProcessorTopology
.
- TopologyBuilder() - Constructor for class org.apache.kafka.streams.processor.TopologyBuilder
-
Create a new builder.
- TopologyBuilder.AutoOffsetReset - Enum in org.apache.kafka.streams.processor
-
Enum used to define auto offset reset policy when creating
KStream
or
KTable
- TopologyBuilder.TopicsInfo - Class in org.apache.kafka.streams.processor
-
- TopologyBuilderException - Exception in org.apache.kafka.streams.errors
-
Indicates a pre-run time error incurred while parsing the
builder
to construct the
processor topology
.
- TopologyBuilderException(String) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- TopologyBuilderException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- TopologyBuilderException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- toStream() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Convert this changelog stream to a
KStream
.
- toStream(KeyValueMapper<? super K, ? super V, ? extends KR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- toString() - Method in class org.apache.kafka.streams.KafkaStreams
-
Produce a string representation containing useful information about this
KafkaStream
instance such as
thread IDs, task IDs, and a representation of the topology DAG including
StateStore
s (cf.
- toString(String) - Method in class org.apache.kafka.streams.KafkaStreams
-
Produce a string representation containing useful information about this
KafkaStream
instance such as
thread IDs, task IDs, and a representation of the topology DAG including
StateStore
s (cf.
- toString() - Method in class org.apache.kafka.streams.KeyValue
-
- toString() - Method in class org.apache.kafka.streams.kstream.Window
-
- toString() - Method in class org.apache.kafka.streams.kstream.Windowed
-
- toString() - Method in class org.apache.kafka.streams.processor.TaskId
-
- toString() - Method in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- toString() - Method in class org.apache.kafka.streams.state.HostInfo
-
- toString() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- transform(TransformerSupplier<? super K, ? super V, KeyValue<K1, V1>>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into zero or more records in the output stream (both key and value type
can be altered arbitrarily).
- transform(K, V) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Transform the record with the given key and value.
- transform(V) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Transform the given value to a new value.
- Transformer<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
The Transformer
interface for stateful mapping of an input record to zero, one, or multiple new output
records (both key and value type can be altered arbitrarily).
- TransformerSupplier<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
A
TransformerSupplier
interface which can create one or more
Transformer
instances.
- transformValues(ValueTransformerSupplier<? super V, ? extends VR>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform the value of each input record into a new value (with possible new type) of the output record.
- WallclockTimestampExtractor - Class in org.apache.kafka.streams.processor
-
Retrieves current wall clock timestamps as System.currentTimeMillis()
.
- WallclockTimestampExtractor() - Constructor for class org.apache.kafka.streams.processor.WallclockTimestampExtractor
-
- Window - Class in org.apache.kafka.streams.kstream
-
A single window instance, defined by its start and end timestamp.
- Window(long, long) - Constructor for class org.apache.kafka.streams.kstream.Window
-
Create a new window for the given start and end time.
- window() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the window containing the values associated with this key.
- WINDOW_STORE_CHANGE_LOG_ADDITIONAL_RETENTION_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
windowstore.changelog.additional.retention.ms
- Windowed<K> - Class in org.apache.kafka.streams.kstream
-
The result key type of a windowed stream aggregation.
- Windowed(K, Window) - Constructor for class org.apache.kafka.streams.kstream.Windowed
-
- windowed(long, long, int, boolean) - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Set the persistent store as a windowed key-value store
- Windows<W extends Window> - Class in org.apache.kafka.streams.kstream
-
The window specification interface for fixed size windows that is used to define window boundaries and window
maintain duration.
- Windows() - Constructor for class org.apache.kafka.streams.kstream.Windows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Not supported by JoinWindows
.
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.Windows
-
Create all windows that contain the provided timestamp, indexed by non-negative window start timestamps.
- windowStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- WindowStore<K,V> - Interface in org.apache.kafka.streams.state
-
- WindowStoreIterator<V> - Interface in org.apache.kafka.streams.state
-
- with(long) - Static method in class org.apache.kafka.streams.kstream.SessionWindows
-
Create a new window specification with the specified inactivity gap in milliseconds.
- withBuiltinTypes(String, Class<K>, Class<V>) - Static method in class org.apache.kafka.streams.state.StateSerdes
-
Create a new instance of
StateSerdes
for the given state name and key-/value-type classes.
- withByteArrayKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be byte arrays.
- withByteArrayValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use byte arrays for values.
- withByteBufferKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be
ByteBuffer
.
- withByteBufferValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use ByteBuffer
for values.
- withDoubleKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be
Double
s.
- withDoubleValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use Double
values.
- withIntegerKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be
Integer
s.
- withIntegerValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use Integer
values.
- withKeys(Class<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withKeys(Serde<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the serializer and deserializer for the keys.
- withLongKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be
Long
s.
- withLongValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use Long
values.
- withStringKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be
String
s.
- withStringValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use String
values.
- withValues(Class<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use values of the specified type.
- withValues(Serde<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use the specified serializer and deserializer for the values.
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- writeAsText(String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- writeAsText(String, String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- writeTo(DataOutputStream) - Method in class org.apache.kafka.streams.processor.TaskId
-
- writeTo(ByteBuffer) - Method in class org.apache.kafka.streams.processor.TaskId
-