All Classes
-
All Classes Interface Summary Class Summary Enum Summary Exception Summary Class Description AdminUtils Utility methods for Kafka admin operations.AssignmentStatus status of partition assignment.Backchannel A backchannel for communication between the commiter -> writer.BackchannelFactory Creates and manages backchannels for the Kafka sink.BackchannelImpl<T> A backchannel for communication between the Kafka committer -> writer.BoundedMode End modes for the Kafka Consumer.CheckpointTransaction An immutable class that represents a transactional id and a checkpoint id.ClusterMetadata ClusterMetadata
provides readers information about a cluster on what topics to read and how to connect to a cluster.DefaultKafkaDatasetFacet Default implementation ofKafkaDatasetFacet
.DefaultKafkaDatasetIdentifier Default implementation ofKafkaDatasetIdentifier
.DefaultKafkaSinkContext Context providing information to assist constructing aProducerRecord
.DefaultTypeDatasetFacet Default implementation ofKafkaDatasetFacet
.DynamicKafkaSource<T> Factory class for the DynamicKafkaSource components.DynamicKafkaSourceBuilder<T> A builder class to make it easier for users to construct aDynamicKafkaSource
.DynamicKafkaSourceEnumerator This enumerator manages multipleKafkaSourceEnumerator
's, which does not have any synchronization since it assumes single threaded execution.DynamicKafkaSourceEnumState The enumerator state keeps track of the state of the sub enumerators assigned splits and metadata.DynamicKafkaSourceEnumStateSerializer (De)serializer forDynamicKafkaSourceEnumState
.DynamicKafkaSourceOptions The connector options forDynamicKafkaSource
that can be passed through the source properties e.g.DynamicKafkaSourceReader<T> Manages state about underlyingKafkaSourceReader
to collect records and commit offsets from multiple Kafka clusters.DynamicKafkaSourceSplit Split that wrapsKafkaPartitionSplit
with Kafka cluster information.DynamicKafkaSourceSplitSerializer (De)serializes theDynamicKafkaSourceSplit
.FlinkFixedPartitioner<T> A partitioner ensuring that each internal Flink partition ends up in one Kafka partition.FlinkKafkaInternalProducer<K,V> AKafkaProducer
that exposes private fields to allow resume producing from a given state.GetMetadataUpdateEvent Event to signal to enumerator that a reader needs to know the current metadata.HeaderProvider<IN> Creates anIterable
ofHeader
s from the input element.JacksonMapperFactory Factory for Jackson mappers.JSONKeyValueDeserializationSchema DeserializationSchema that deserializes a JSON String into an ObjectNode.KafkaClusterMetricGroup A custom proxy metric group in order to groupKafkaSourceReaderMetrics
by Kafka cluster.KafkaClusterMetricGroupManager Manages metric groups for each cluster.KafkaCommittable This class holds the necessary information to construct a newFlinkKafkaInternalProducer
to commit transactions inKafkaCommitter
.KafkaCommitter Committer implementation forKafkaSink
KafkaConnectorOptions Options for the Kafka connector.KafkaConnectorOptions.ScanBoundedMode Bounded mode for the Kafka consumer, seeKafkaConnectorOptions.SCAN_BOUNDED_MODE
.KafkaConnectorOptions.ScanStartupMode Startup mode for the Kafka consumer, seeKafkaConnectorOptions.SCAN_STARTUP_MODE
.KafkaConnectorOptions.ValueFieldsStrategy Strategies to derive the data type of a value format by considering a key format.KafkaDatasetFacet Facet definition to contain all Kafka specific information on Kafka sources and sinks.KafkaDatasetFacetProvider Contains method to extractKafkaDatasetFacet
.KafkaDatasetIdentifier Kafka dataset identifier which can contain either a list of topics or a topic pattern.KafkaDatasetIdentifierProvider Contains method which allows extracting topic identifier.KafkaDynamicSink A version-agnostic KafkaDynamicTableSink
.KafkaDynamicSource A version-agnostic KafkaScanTableSource
.KafkaDynamicTableFactory Factory for creating configured instances ofKafkaDynamicSource
andKafkaDynamicSink
.KafkaMetadataService Metadata service that returns Kafka details.KafkaMetricMutableWrapper Gauge for getting the current value of a Kafka metric.KafkaPartitioner<T> AKafkaPartitioner
wraps logic on how to partition records across partitions of multiple Kafka topics.KafkaPartitionSplit ASourceSplit
for a Kafka partition.KafkaPartitionSplitReader ASplitReader
implementation that reads records from Kafka partitions.KafkaPartitionSplitReaderWrapper This extends to Kafka Partition Split Reader to wrap split ids with the cluster name.KafkaPartitionSplitSerializer Theserializer
forKafkaPartitionSplit
.KafkaPartitionSplitState This class extends KafkaPartitionSplit to track a mutable current offset.KafkaPropertiesUtil Utility class for modify Kafka properties.KafkaRecordDeserializationSchema<T> An interface for the deserialization of Kafka records.KafkaRecordEmitter<T> TheRecordEmitter
implementation forKafkaSourceReader
.KafkaRecordSerializationSchema<T> A serialization schema which defines how to convert a value of typeT
toProducerRecord
.KafkaRecordSerializationSchema.KafkaSinkContext Context providing information of the kafka record target location.KafkaRecordSerializationSchemaBuilder<IN> Builder to constructKafkaRecordSerializationSchema
.KafkaSink<IN> Flink Sink to produce data into a Kafka topic.KafkaSinkBuilder<IN> Builder to constructKafkaSink
.KafkaSource<OUT> The Source implementation of Kafka.KafkaSourceBuilder<OUT> The builder class forKafkaSource
to make it easier for the users to construct aKafkaSource
.KafkaSourceEnumerator The enumerator class for Kafka source.KafkaSourceEnumerator.PartitionOffsetsRetrieverImpl The implementation for offsets retriever with a consumer and an admin client.KafkaSourceEnumState The state of Kafka source enumerator.KafkaSourceEnumStateSerializer TheSerializer
for the enumerator state of Kafka source.KafkaSourceFetcherManager The SplitFetcherManager for Kafka source.KafkaSourceOptions Configurations for KafkaSource.KafkaSourceReader<T> The source reader for Kafka partitions.KafkaSourceReaderMetrics A collection class for handling metrics inKafkaSourceReader
.KafkaStream Kafka stream represents multiple topics over multiple Kafka clusters and this class encapsulates all the necessary information to initiate Kafka consumers to read a stream.KafkaStreamSetSubscriber Subscribe to streams based on the set of ids.KafkaStreamSubscriber The subscriber interacts withKafkaMetadataService
to find which Kafka streams the source will subscribe to.KafkaSubscriber Kafka consumer allows a few different ways to consume from the topics, including: Subscribe from a collection of topics.KafkaTopicPartitionStateSentinel Magic values used to represent special offset states before partitions are actually read.KafkaWriterState The state of the Kafka writer.LineageUtil Utility class with useful methods for managing lineage objects.MetadataUpdateEvent SignalsDynamicKafkaSourceReader
to stop their underlying readers.MetricUtil Collection of methods to interact with Kafka's client metric system.NoopCommitter The committer to be used for non exactly-once delivery guarantees.NoStoppingOffsetsInitializer An implementation ofOffsetsInitializer
which does not initialize anything.OffsetsInitializer An interface for users to specify the starting / stopping offset of aKafkaPartitionSplit
.OffsetsInitializer.PartitionOffsetsRetriever An interface that provides necessary information to theOffsetsInitializer
to get the initial offsets of the Kafka partitions.OffsetsInitializerValidator Interface for validatingOffsetsInitializer
with properties fromKafkaSource
.ProducerPool A pool of producers that can be recycled.ProducerPoolImpl Manages a pool ofFlinkKafkaInternalProducer
instances for reuse in theKafkaWriter
and keeps track of the used transactional ids.ReadableBackchannel<T> The readable portion of a backchannel for communication between the commiter -> writer.SingleClusterTopicMetadataService AKafkaMetadataService
that delegates metadata fetching to a singleAdminClient
, which is scoped to a single cluster.SinkBufferFlushMode Sink buffer flush configuration.StartupMode Startup modes for the Kafka Consumer.StoppableKafkaEnumContextProxy A proxy enumerator context that supports life cycle management of underlying threads related to a subKafkaSourceEnumerator
.StoppableKafkaEnumContextProxy.HandledFlinkKafkaException General exception to signal to internal exception handling mechanisms that a benign error occurred.StoppableKafkaEnumContextProxy.StoppableKafkaEnumContextProxyFactory This factory exposes a way to override theStoppableKafkaEnumContextProxy
used in the enumerator.StreamPatternSubscriber To subscribe to streams based on a pattern.TopicPartitionAndAssignmentStatus Kafka partition with assign status.TopicSelector<IN> Selects a topic for the incoming record.TransactionAbortStrategyContextImpl Implementation ofTransactionAbortStrategyImpl.Context
.TransactionAbortStrategyImpl Implementations of an abort strategy for transactions left over from previous runs.TransactionAbortStrategyImpl.Context Context for theTransactionAbortStrategyImpl
.TransactionAbortStrategyImpl.TransactionAborter Injects the actual abortion of the transactional id generated by one of the strategies.TransactionalIdFactory Utility class for constructing transactionalIds for Kafka transactions.TransactionFinished Represents the end of a transaction.TransactionNamingStrategy The strategy to name transactions.TransactionNamingStrategyContextImpl Implementation ofTransactionNamingStrategyImpl.Context
.TransactionNamingStrategyImpl Implementation ofTransactionNamingStrategy
.TransactionNamingStrategyImpl.Context Context for the transaction naming strategy.TransactionOwnership Describes the ownership model of transactional ids and with that ownership of the transactions.TwoPhaseCommittingStatefulSink<InputT,WriterStateT,CommT> A combination ofSupportsCommitter
andSupportsWriterState
.TwoPhaseCommittingStatefulSink.PrecommittingStatefulSinkWriter<InputT,WriterStateT,CommT> A combination ofStatefulSinkWriter
.TypeDatasetFacet Facet definition to contain type information of source and sink.TypeDatasetFacetProvider Contains method to extractTypeDatasetFacet
.UpsertKafkaDynamicTableFactory Upsert-Kafka factory.UpsertKafkaDynamicTableFactory.DecodingFormatWrapper It is used to wrap the decoding format and expose the desired changelog mode.UpsertKafkaDynamicTableFactory.EncodingFormatWrapper It is used to wrap the encoding format and expose the desired changelog mode.WritableBackchannel<T> The writable portion of aBackchannel
for communication between the commiter -> writer.