Skip navigation links
A C E F G I K O P S W 

A

assignPartitions(KafkaConsumer<?, ?>, List<TopicPartition>) - Method in class org.apache.flink.streaming.connectors.kafka.internal.KafkaConsumerCallBridge010
 

C

createCallBridge() - Method in class org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher
This method needs to be overridden because Kafka broke binary compatibility between 0.9 and 0.10, changing binary signatures
createFetcher(SourceFunction.SourceContext<T>, Map<KafkaTopicPartition, Long>, SerializedValue<AssignerWithPeriodicWatermarks<T>>, SerializedValue<AssignerWithPunctuatedWatermarks<T>>, StreamingRuntimeContext, OffsetCommitMode) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer010
 

E

emitRecord(T, KafkaTopicPartitionState<TopicPartition>, long, ConsumerRecord<?, ?>) - Method in class org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher
 

F

FlinkKafkaConsumer010<T> - Class in org.apache.flink.streaming.connectors.kafka
The Flink Kafka Consumer is a streaming data source that pulls a parallel data stream from Apache Kafka 0.10.x.
FlinkKafkaConsumer010(String, DeserializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer010
Creates a new Kafka streaming source consumer for Kafka 0.10.x
FlinkKafkaConsumer010(String, KeyedDeserializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer010
Creates a new Kafka streaming source consumer for Kafka 0.10.x This constructor allows passing a KeyedDeserializationSchema for reading key/value pairs, offsets, and topic names from Kafka.
FlinkKafkaConsumer010(List<String>, DeserializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer010
Creates a new Kafka streaming source consumer for Kafka 0.10.x This constructor allows passing multiple topics to the consumer.
FlinkKafkaConsumer010(List<String>, KeyedDeserializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer010
Creates a new Kafka streaming source consumer for Kafka 0.10.x This constructor allows passing multiple topics and a key/value deserialization schema.
FlinkKafkaProducer010<T> - Class in org.apache.flink.streaming.connectors.kafka
Flink Sink to produce data into a Kafka topic.
FlinkKafkaProducer010(String, String, SerializationSchema<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
FlinkKafkaProducer010(String, SerializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
FlinkKafkaProducer010(String, SerializationSchema<T>, Properties, FlinkKafkaPartitioner<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
FlinkKafkaProducer010(String, String, KeyedSerializationSchema<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
FlinkKafkaProducer010(String, KeyedSerializationSchema<T>, Properties) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
FlinkKafkaProducer010(String, KeyedSerializationSchema<T>, Properties, FlinkKafkaPartitioner<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Create Kafka producer This constructor does not allow writing timestamps to Kafka, it follow approach (a) (see above)
FlinkKafkaProducer010(String, SerializationSchema<T>, Properties, KafkaPartitioner<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Deprecated.
This is a deprecated since it does not correctly handle partitioning when producing to multiple topics. Use FlinkKafkaProducer010.FlinkKafkaProducer010(String, SerializationSchema, Properties, FlinkKafkaPartitioner) instead.
FlinkKafkaProducer010(String, KeyedSerializationSchema<T>, Properties, KafkaPartitioner<T>) - Constructor for class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Deprecated.
This is a deprecated constructor that does not correctly handle partitioning when producing to multiple topics. Use FlinkKafkaProducer010.FlinkKafkaProducer010(String, SerializationSchema, Properties, FlinkKafkaPartitioner) instead.
FlinkKafkaProducer010.FlinkKafkaProducer010Configuration<T> - Class in org.apache.flink.streaming.connectors.kafka
Configuration object returned by the writeToKafkaWithTimestamps() call.

G

getFetcherName() - Method in class org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher
 
getIterationRuntimeContext() - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
This method is used for approach (a) (see above)

I

initializeState(FunctionInitializationContext) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
 
invoke(T) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Invoke method for using the Sink as DataStream.addSink() sink.

K

Kafka010AvroTableSource - Class in org.apache.flink.streaming.connectors.kafka
Kafka StreamTableSource for Kafka 0.10.
Kafka010AvroTableSource(String, Properties, Class<? extends SpecificRecordBase>) - Constructor for class org.apache.flink.streaming.connectors.kafka.Kafka010AvroTableSource
Creates a Kafka 0.10 Avro StreamTableSource using a given SpecificRecord.
Kafka010Fetcher<T> - Class in org.apache.flink.streaming.connectors.kafka.internal
A fetcher that fetches data from Kafka brokers via the Kafka 0.10 consumer API.
Kafka010Fetcher(SourceFunction.SourceContext<T>, Map<KafkaTopicPartition, Long>, SerializedValue<AssignerWithPeriodicWatermarks<T>>, SerializedValue<AssignerWithPunctuatedWatermarks<T>>, ProcessingTimeService, long, ClassLoader, String, MetricGroup, KeyedDeserializationSchema<T>, Properties, long, boolean) - Constructor for class org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher
 
Kafka010JsonTableSource - Class in org.apache.flink.streaming.connectors.kafka
Kafka StreamTableSource for Kafka 0.10.
Kafka010JsonTableSource(String, Properties, TypeInformation<Row>) - Constructor for class org.apache.flink.streaming.connectors.kafka.Kafka010JsonTableSource
Creates a Kafka 0.10 JSON StreamTableSource.
Kafka010TableSource - Class in org.apache.flink.streaming.connectors.kafka
Kafka StreamTableSource for Kafka 0.10.
Kafka010TableSource(String, Properties, DeserializationSchema<Row>, TypeInformation<Row>) - Constructor for class org.apache.flink.streaming.connectors.kafka.Kafka010TableSource
Creates a Kafka 0.10 StreamTableSource.
KafkaConsumerCallBridge010 - Class in org.apache.flink.streaming.connectors.kafka.internal
The ConsumerCallBridge simply calls the KafkaConsumer.assign(java.util.Collection) method.
KafkaConsumerCallBridge010() - Constructor for class org.apache.flink.streaming.connectors.kafka.internal.KafkaConsumerCallBridge010
 

O

open(Configuration) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
This method is used for approach (a) (see above)
org.apache.flink.streaming.connectors.kafka - package org.apache.flink.streaming.connectors.kafka
 
org.apache.flink.streaming.connectors.kafka.internal - package org.apache.flink.streaming.connectors.kafka.internal
 

P

processElement(StreamRecord<T>) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Process method for using the sink with timestamp support.

S

seekPartitionToBeginning(KafkaConsumer<?, ?>, TopicPartition) - Method in class org.apache.flink.streaming.connectors.kafka.internal.KafkaConsumerCallBridge010
 
seekPartitionToEnd(KafkaConsumer<?, ?>, TopicPartition) - Method in class org.apache.flink.streaming.connectors.kafka.internal.KafkaConsumerCallBridge010
 
setFlushOnCheckpoint(boolean) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010.FlinkKafkaProducer010Configuration
If set to true, the Flink producer will wait for all outstanding messages in the Kafka buffers to be acknowledged by the Kafka producer on a checkpoint.
setFlushOnCheckpoint(boolean) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
If set to true, the Flink producer will wait for all outstanding messages in the Kafka buffers to be acknowledged by the Kafka producer on a checkpoint.
setLogFailuresOnly(boolean) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010.FlinkKafkaProducer010Configuration
Defines whether the producer should fail on errors, or only log them.
setLogFailuresOnly(boolean) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Defines whether the producer should fail on errors, or only log them.
setRuntimeContext(RuntimeContext) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
This method is used for approach (a) (see above)
setWriteTimestampToKafka(boolean) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010.FlinkKafkaProducer010Configuration
If set to true, Flink will write the (event time) timestamp attached to each record into Kafka.
snapshotState(FunctionSnapshotContext) - Method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
 

W

writeToKafkaWithTimestamps(DataStream<T>, String, KeyedSerializationSchema<T>, Properties) - Static method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
writeToKafkaWithTimestamps(DataStream<T>, String, SerializationSchema<T>, Properties) - Static method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
writeToKafkaWithTimestamps(DataStream<T>, String, KeyedSerializationSchema<T>, Properties, FlinkKafkaPartitioner<T>) - Static method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Creates a FlinkKafkaProducer for a given topic.
writeToKafkaWithTimestamps(DataStream<T>, String, KeyedSerializationSchema<T>, Properties, KafkaPartitioner<T>) - Static method in class org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010
Deprecated.
This is a deprecated since it does not correctly handle partitioning when producing to multiple topics. Use FlinkKafkaProducer010.FlinkKafkaProducer010(String, SerializationSchema, Properties, FlinkKafkaPartitioner) instead.
A C E F G I K O P S W 
Skip navigation links

Copyright © 2014–2017 The Apache Software Foundation. All rights reserved.