http://git-wip-us.apache.org/repos/asf/kafka-site/blob/2e200cfc/10/javadoc/index-all.html ---------------------------------------------------------------------- diff --git a/10/javadoc/index-all.html b/10/javadoc/index-all.html new file mode 100644 index 0000000..cde330c --- /dev/null +++ b/10/javadoc/index-all.html @@ -0,0 +1,10050 @@ + + + + + +Index (kafka 1.0.0 API) + + + + + + + + +
+ + +
Skip navigation links
+ + + + +
+ + +
A B C D E F G H I J K L M N O P Q R S T U V W Z  + + +

A

+
+
abortTransaction() - Method in class org.apache.kafka.clients.producer.KafkaProducer
+
+
Aborts the ongoing transaction.
+
+
abortTransaction() - Method in class org.apache.kafka.clients.producer.MockProducer
+
 
+
abortTransaction() - Method in interface org.apache.kafka.clients.producer.Producer
+
+ +
+
AbstractConfig - Class in org.apache.kafka.common.config
+
+
A convenient base class for configurations to extend.
+
+
AbstractConfig(ConfigDef, Map<?, ?>, boolean) - Constructor for class org.apache.kafka.common.config.AbstractConfig
+
 
+
AbstractConfig(ConfigDef, Map<?, ?>) - Constructor for class org.apache.kafka.common.config.AbstractConfig
+
 
+
AbstractNotifyingBatchingRestoreCallback - Class in org.apache.kafka.streams.processor
+
+
Abstract implementation of the BatchingStateRestoreCallback used for batch restoration operations.
+
+
AbstractNotifyingBatchingRestoreCallback() - Constructor for class org.apache.kafka.streams.processor.AbstractNotifyingBatchingRestoreCallback
+
 
+
AbstractNotifyingRestoreCallback - Class in org.apache.kafka.streams.processor
+
+
Abstract implementation of the StateRestoreCallback used for batch restoration operations.
+
+
AbstractNotifyingRestoreCallback() - Constructor for class org.apache.kafka.streams.processor.AbstractNotifyingRestoreCallback
+
 
+
AbstractOptions<T extends AbstractOptions> - Class in org.apache.kafka.clients.admin
+
 
+
AbstractOptions() - Constructor for class org.apache.kafka.clients.admin.AbstractOptions
+
 
+
AbstractProcessor<K,V> - Class in org.apache.kafka.streams.processor
+
+
An abstract implementation of Processor that manages the ProcessorContext instance and provides default no-op + implementations of AbstractProcessor.punctuate(long) and AbstractProcessor.close().
+
+
AbstractProcessor() - Constructor for class org.apache.kafka.streams.processor.AbstractProcessor
+
 
+
accept(A, B) - Method in class org.apache.kafka.common.KafkaFuture.BiConsumer
+
 
+
accepts(StateStore) - Method in interface org.apache.kafka.streams.state.QueryableStoreType
+
+
Called when searching for StateStores to see if they + match the type expected by implementors of this interface
+
+
AccessControlEntry - Class in org.apache.kafka.common.acl
+
+
Represents an access control entry.
+
+
AccessControlEntry(String, String, AclOperation, AclPermissionType) - Constructor for class org.apache.kafka.common.acl.AccessControlEntry
+
+
Create an instance of an access control entry with the provided parameters.
+
+
AccessControlEntryFilter - Class in org.apache.kafka.common.acl
+
+
Represents a filter which matches access control entries.
+
+
AccessControlEntryFilter(String, String, AclOperation, AclPermissionType) - Constructor for class org.apache.kafka.common.acl.AccessControlEntryFilter
+
+
Create an instance of an access control entry filter with the provided parameters.
+
+
ACKS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
+
+
acks
+
+
AclBinding - Class in org.apache.kafka.common.acl
+
+
Represents a binding between a resource and an access control entry.
+
+
AclBinding(Resource, AccessControlEntry) - Constructor for class org.apache.kafka.common.acl.AclBinding
+
+
Create an instance of this class with the provided parameters.
+
+
AclBindingFilter - Class in org.apache.kafka.common.acl
+
+
A filter which can match AclBinding objects.
+
+
AclBindingFilter(ResourceFilter, AccessControlEntryFilter) - Constructor for class org.apache.kafka.common.acl.AclBindingFilter
+
+
Create an instance of this filter with the provided parameters.
+
+
AclOperation - Enum in org.apache.kafka.common.acl
+
+
Represents an operation which an ACL grants or denies permission to perform.
+
+
AclPermissionType - Enum in org.apache.kafka.common.acl
+
+
Represents whether an ACL grants or denies permissions.
+
+
activeTasks() - Method in class org.apache.kafka.streams.processor.ThreadMetadata
+
 
+
add(Header) - Method in interface org.apache.kafka.common.header.Headers
+
+
Adds a header (key inside), to the end, returning if the operation succeeded.
+
+
add(String, byte[]) - Method in interface org.apache.kafka.common.header.Headers
+
+
Creates and adds a header, to the end, returning if the operation succeeded.
+
+
addClientSaslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SaslConfigs
+
 
+
addClientSslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SslConfigs
+
 
+
addDeserializerToConfig(Map<String, Object>, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
+
 
+
addDeserializerToConfig(Properties, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
+
 
+
addErrorMessage(String) - Method in class org.apache.kafka.common.config.ConfigValue
+
 
+
addGlobalStore(StateStoreSupplier<KeyValueStore>, String, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Adds a global StateStore to the topology.
+
+
addGlobalStore(StateStoreSupplier<KeyValueStore>, String, TimestampExtractor, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Adds a global StateStore to the topology.
+
+
addGlobalStore(StoreBuilder, String, String, Consumed, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.StreamsBuilder
+
+
Adds a global StateStore to the topology.
+
+
addGlobalStore(StoreBuilder, String, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.Topology
+
+
Adds a global StateStore to the topology.
+
+
addGlobalStore(StoreBuilder, String, TimestampExtractor, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.Topology
+
+
Adds a global StateStore to the topology.
+
+
addInternalTopic(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Adds an internal topic + + NOTE this function would not needed by developers working with the processor APIs, but only used + for the high-level DSL parsing functionalities.
+
+
addLatencyAndThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
+
+
Add a latency and throughput sensor for a specific operation, which will include the following sensors: + + average latency + max latency + throughput (num.operations / time unit) + + Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the + same scope if it has not been created.
+
+
addProcessor(String, ProcessorSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new processor node that receives and processes records output by one or more predecessor source or processor node.
+
+
addProcessor(String, ProcessorSupplier, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new processor node that receives and processes records output by one or more parent source or processor + node.
+
+
addRecord(ConsumerRecord<K, V>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
+
 
+
addSensor(String, Sensor.RecordingLevel) - Method in interface org.apache.kafka.streams.StreamsMetrics
+
+
Generic method to create a sensor.
+
+
addSensor(String, Sensor.RecordingLevel, Sensor...) - Method in interface org.apache.kafka.streams.StreamsMetrics
+
+
Generic method to create a sensor with parent sensors.
+
+
addSerializerToConfig(Map<String, Object>, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
+
 
+
addSerializerToConfig(Properties, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
+
 
+
addSink(String, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new sink that forwards records from predecessor nodes (processors and/or sources) to the named Kafka topic.
+
+
addSink(String, String, StreamPartitioner, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new sink that forwards records from predecessor nodes (processors and/or sources) to the named Kafka topic, using + the supplied partitioner.
+
+
addSink(String, String, Serializer, Serializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new sink that forwards records from predecessor nodes (processors and/or sources) to the named Kafka topic.
+
+
addSink(String, String, Serializer<K>, Serializer<V>, StreamPartitioner<? super K, ? super V>, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new sink that forwards records from predecessor nodes (processors and/or sources) to the named Kafka topic.
+
+
addSink(String, String, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
+
+
addSink(String, String, StreamPartitioner, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic, + using the supplied partitioner.
+
+
addSink(String, String, Serializer, Serializer, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
+
+
addSink(String, String, Serializer<K>, Serializer<V>, StreamPartitioner<? super K, ? super V>, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
+
+
addSource(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
+
+
addSource(String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forwards the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forwards the records to child processor and/or sink nodes.
+
+
addSource(TopologyBuilder.AutoOffsetReset, String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Add a new source that consumes from topics matching the given pattern + and forwards the records to child processor and/or sink nodes.
+
+
addSource(String, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(String, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern + and forward the records to child processor and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, TimestampExtractor, String, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, TimestampExtractor, String, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern and forward the records to child processor + and/or sink nodes.
+
+
addSource(String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
+
+
addSource(String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern and forwards the records to child processor + and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern and forwards the records to child processor + and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern and forwards the records to child processor + and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
+
+
addSource(Topology.AutoOffsetReset, String, TimestampExtractor, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.Topology
+
+
Add a new source that consumes from topics matching the given pattern and forwards the records to child processor + and/or sink nodes.
+
+
addStateStore(StateStoreSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
+
+
Deprecated.
+
Adds a state store
+
+
addStateStore(StoreBuilder) - Method in class org.apache.kafka.streams.StreamsBuilder
+
+
Adds a state store to the underlying Topology.
+
+
addStateStore(StoreBuilder, String...) - Method in class org.apache.kafka.streams.Topology
+
+
Adds a state store.
+
+
addThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
+
+
Add a throughput sensor for a specific operation: + + throughput (num.operations / time unit) + + Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the + same scope if it has not been created.
+
+
addWaiter(KafkaFuture.BiConsumer<? super T, ? super Throwable>) - Method in class org.apache.kafka.common.KafkaFuture
+
 
+
AdminClient - Class in org.apache.kafka.clients.admin
+
+
The administrative client for Kafka, which supports managing and inspecting topics, brokers, configurations and ACLs.
+
+
AdminClient() - Constructor for class org.apache.kafka.clients.admin.AdminClient
+
 
+
AdminClientConfig - Class in org.apache.kafka.clients.admin
+
+
The AdminClient configuration class, which also contains constants for configuration entry names.
+
+
AdminClientConfig(Map<?, ?>) - Constructor for class org.apache.kafka.clients.admin.AdminClientConfig
+
 
+
advanceBy(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
+
+
Return a window definition with the original size, but advance ("hop") the window by the given interval, which + specifies by how much a window moves forward relative to the previous one.
+
+
advanceMs - Variable in class org.apache.kafka.streams.kstream.TimeWindows
+
+
The size of the window's advance interval in milliseconds, i.e., by how much a window moves forward relative to + the previous one.
+
+
after(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
+
+
Changes the end window boundary to timeDifferenceMs but keep the start window boundary as is.
+
+
afterMs - Variable in class org.apache.kafka.streams.kstream.JoinWindows
+
+
Maximum time difference for tuples that are after the join tuple.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Materialized<K, VR, KeyValueStore<Bytes, byte[]>>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+
Aggregate the values of records in this stream by the grouped key.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+
Aggregate the values of records in this stream by the grouped key.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+
Deprecated. +
use windowedBy(windows) followed by + TimeWindowedKStream.aggregate(Initializer, Aggregator, Materialized) aggregate(initializer, aggregator, Materialized.with(null, aggValueSerde))}
+
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Materialized<K, VR, KeyValueStore<Bytes, byte[]>>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+
Aggregate the value of records of the original KTable that got mapped to the same key into a new instance of KTable using default serializers and deserializers.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+
Aggregate the value of records of the original KTable that got mapped to the same key into a new instance of KTable using default serializers and deserializers.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
+
+ +
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Merger<? super K, VR>) - Method in interface org.apache.kafka.streams.kstream.SessionWindowedKStream
+
+
Aggregate the values of records in this stream by the grouped key and defined SessionWindows.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Merger<? super K, VR>, Materialized<K, VR, SessionStore<Bytes, byte[]>>) - Method in interface org.apache.kafka.streams.kstream.SessionWindowedKStream
+
+
Aggregate the values of records in this stream by the grouped key and defined SessionWindows.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>) - Method in interface org.apache.kafka.streams.kstream.TimeWindowedKStream
+
+
Aggregate the values of records in this stream by the grouped key.
+
+
aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Materialized<K, VR, WindowStore<Bytes, byte[]>>) - Method in interface org.apache.kafka.streams.kstream.TimeWindowedKStream
+
+
Aggregate the values of records in this stream by the grouped key.
+
+
Aggregator<K,V,VA> - Interface in org.apache.kafka.streams.kstream
+
+
The Aggregator interface for aggregating values of the given key.
+
+
all() - Method in class org.apache.kafka.clients.admin.AlterConfigsResult
+
+
Return a future which succeeds only if all the alter configs operations succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.AlterReplicaLogDirsResult
+
+
Return a future which succeeds if all the replica movement have succeeded
+
+
all() - Method in class org.apache.kafka.clients.admin.CreateAclsResult
+
+
Return a future which succeeds only if all the ACL creations succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.CreatePartitionsResult
+
+
Return a future which succeeds if all the partition creations succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.CreateTopicsResult
+
+
Return a future which succeeds if all the topic creations succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.DeleteAclsResult
+
+
Return a future which succeeds only if all the ACLs deletions succeed, and which contains all the deleted ACLs.
+
+
all() - Method in class org.apache.kafka.clients.admin.DeleteTopicsResult
+
+
Return a future which succeeds only if all the topic deletions succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.DescribeConfigsResult
+
+
Return a future which succeeds only if all the config descriptions succeed.
+
+
all() - Method in class org.apache.kafka.clients.admin.DescribeLogDirsResult
+
+
Return a future which succeeds only if all the brokers have responded without error
+
+
all() - Method in class org.apache.kafka.clients.admin.DescribeReplicaLogDirsResult
+
+
Return a future which succeeds if log directory information of all replicas are available
+
+
all() - Method in class org.apache.kafka.clients.admin.DescribeTopicsResult
+
+
Return a future which succeeds only if all the topic descriptions succeed.
+
+
all() - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
+
+
Return an iterator over all keys in this store.
+
+
allMetadata() - Method in class org.apache.kafka.streams.KafkaStreams
+
+
Find all currently running KafkaStreams instances (potentially remotely) that use the same + application ID as this instance (i.e., all instances that belong to + the same Kafka Streams application) and return StreamsMetadata for each discovered instance.
+
+
allMetadataForStore(String) - Method in class org