Modifier and Type | Class and Description |
---|---|
class |
AbstractDeserializationSchema<T>
The deserialization schema describes how to turn the byte messages delivered by certain data
sources (for example Apache Kafka) into data types (Java/Scala objects) that are processed by
Flink.
|
class |
SimpleStringSchema
Very simple serialization schema for strings.
|
class |
TypeInformationSerializationSchema<T>
A serialization and deserialization schema that uses Flink's serialization stack to transform
typed from and to byte arrays.
|
Modifier and Type | Class and Description |
---|---|
class |
AvroDeserializationSchema<T>
Deserialization schema that deserializes from Avro binary format.
|
class |
AvroRowDataDeserializationSchema
Deserialization schema from Avro bytes to
RowData . |
class |
AvroRowDeserializationSchema
Deserialization schema from Avro bytes to
Row . |
class |
RegistryAvroDeserializationSchema<T>
Deserialization schema that deserializes from Avro binary format using
SchemaCoder . |
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
AvroRowFormatFactory.createDeserializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
AvroFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
ConfluentRegistryAvroDeserializationSchema<T>
Deserialization schema that deserializes from Avro binary format using
SchemaCoder that
uses Confluent Schema Registry. |
Modifier and Type | Class and Description |
---|---|
class |
CsvRowDataDeserializationSchema
Deserialization schema from CSV to Flink Table & SQL internal data structures.
|
class |
CsvRowDeserializationSchema
Deserialization schema from CSV to Flink types.
|
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
CsvRowFormatFactory.createDeserializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
CsvFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
JsonNodeDeserializationSchema
DeserializationSchema that deserializes a JSON String into an ObjectNode.
|
class |
JsonRowDataDeserializationSchema
Deserialization schema from JSON to Flink Table/SQL internal data structure
RowData . |
class |
JsonRowDeserializationSchema
Deserialization schema from JSON to Flink types.
|
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
JsonRowFormatFactory.createDeserializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
JsonFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
CanalJsonDeserializationSchema
Deserialization schema from Canal JSON to Flink Table/SQL internal data structure
RowData . |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
CanalJsonFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
DebeziumJsonDeserializationSchema
Deserialization schema from Debezium JSON to Flink Table/SQL internal data structure
RowData . |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
DebeziumJsonFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Method and Description |
---|---|
<OUT> PubSubSource.ProjectNameBuilder<OUT> |
PubSubSource.DeserializationSchemaBuilder.withDeserializationSchema(DeserializationSchema<OUT> deserializationSchema)
Set the DeserializationSchema used to deserialize incoming PubSubMessages.
|
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
KafkaTableSourceBase.getDeserializationSchema()
Returns the deserialization schema.
|
Modifier and Type | Method and Description |
---|---|
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka011TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka010TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected abstract FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a version-specific Kafka consumer.
|
protected KafkaTableSourceBase |
KafkaTableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected KafkaTableSourceBase |
Kafka011TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected KafkaTableSourceBase |
Kafka010TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected abstract KafkaTableSourceBase |
KafkaTableSourceSinkFactoryBase.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Constructs the version-specific Kafka table source.
|
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.getKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Returns a version-specific Kafka consumer with the start position configured.
|
Constructor and Description |
---|
FlinkKafkaConsumer(List<String> topics,
DeserializationSchema<T> deserializer,
Properties props)
Creates a new Kafka streaming source consumer.
|
FlinkKafkaConsumer(Pattern subscriptionPattern,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer.
|
FlinkKafkaConsumer(String topic,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer.
|
FlinkKafkaConsumer010(List<String> topics,
DeserializationSchema<T> deserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.10.x
|
FlinkKafkaConsumer010(Pattern subscriptionPattern,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.10.x.
|
FlinkKafkaConsumer010(String topic,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.10.x.
|
FlinkKafkaConsumer011(List<String> topics,
DeserializationSchema<T> deserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.11.x
|
FlinkKafkaConsumer011(Pattern subscriptionPattern,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.11.x.
|
FlinkKafkaConsumer011(String topic,
DeserializationSchema<T> valueDeserializer,
Properties props)
Creates a new Kafka streaming source consumer for Kafka 0.11.x.
|
Kafka010TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a Kafka 0.10
StreamTableSource . |
Kafka010TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.10
StreamTableSource . |
Kafka011TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a Kafka 0.11
StreamTableSource . |
Kafka011TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.11
StreamTableSource . |
KafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSourceBase(TableSchema schema,
Optional<String> proctimeAttribute,
List<RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSourceBase(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a generic Kafka
StreamTableSource . |
Constructor and Description |
---|
KafkaDeserializationSchemaWrapper(DeserializationSchema<T> deserializationSchema) |
Modifier and Type | Field and Description |
---|---|
protected DecodingFormat<DeserializationSchema<RowData>> |
KafkaDynamicSourceBase.decodingFormat
Scan format for decoding records from Kafka.
|
Modifier and Type | Method and Description |
---|---|
protected FlinkKafkaConsumerBase<RowData> |
KafkaDynamicSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<RowData> deserializationSchema) |
protected FlinkKafkaConsumerBase<RowData> |
Kafka011DynamicSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<RowData> deserializationSchema) |
protected FlinkKafkaConsumerBase<RowData> |
Kafka010DynamicSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<RowData> deserializationSchema) |
protected abstract FlinkKafkaConsumerBase<RowData> |
KafkaDynamicSourceBase.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<RowData> deserializationSchema)
Creates a version-specific Kafka consumer.
|
protected FlinkKafkaConsumerBase<RowData> |
KafkaDynamicSourceBase.getKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<RowData> deserializationSchema)
Returns a version-specific Kafka consumer with the start position configured.
|
Modifier and Type | Method and Description |
---|---|
protected KafkaDynamicSourceBase |
KafkaDynamicTableFactory.createKafkaTableSource(DataType producedDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected KafkaDynamicSourceBase |
Kafka011DynamicTableFactory.createKafkaTableSource(DataType producedDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected KafkaDynamicSourceBase |
Kafka010DynamicTableFactory.createKafkaTableSource(DataType producedDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis) |
protected abstract KafkaDynamicSourceBase |
KafkaDynamicTableFactoryBase.createKafkaTableSource(DataType producedDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Constructs the version-specific Kafka table source.
|
Constructor and Description |
---|
Kafka010DynamicSource(DataType outputDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a Kafka 0.10
StreamTableSource . |
Kafka011DynamicSource(DataType outputDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a Kafka 0.11
ScanTableSource . |
KafkaDynamicSource(DataType outputDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a generic Kafka
StreamTableSource . |
KafkaDynamicSourceBase(DataType outputDataType,
String topic,
Properties properties,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis)
Creates a generic Kafka
StreamTableSource . |
Constructor and Description |
---|
FlinkDynamoDBStreamsConsumer(String stream,
DeserializationSchema<T> deserializer,
Properties config)
Constructor of FlinkDynamoDBStreamsConsumer.
|
FlinkKinesisConsumer(String stream,
DeserializationSchema<T> deserializer,
Properties configProps)
Creates a new Flink Kinesis Consumer.
|
Constructor and Description |
---|
KinesisDeserializationSchemaWrapper(DeserializationSchema<T> deserializationSchema) |
Modifier and Type | Field and Description |
---|---|
protected DeserializationSchema<OUT> |
RMQSource.schema |
Constructor and Description |
---|
RMQSource(RMQConnectionConfig rmqConnectionConfig,
String queueName,
boolean usesCorrelationId,
DeserializationSchema<OUT> deserializationSchema)
Creates a new RabbitMQ source.
|
RMQSource(RMQConnectionConfig rmqConnectionConfig,
String queueName,
DeserializationSchema<OUT> deserializationSchema)
Creates a new RabbitMQ source with at-least-once message processing guarantee when
checkpointing is enabled.
|
Modifier and Type | Class and Description |
---|---|
class |
EventDeSerializer
A serializer and deserializer for the
Event type. |
Modifier and Type | Class and Description |
---|---|
class |
KafkaEventSchema
The serialization schema for the
KafkaEvent type. |
Modifier and Type | Interface and Description |
---|---|
interface |
DeserializationSchema<T>
Deprecated.
Use
DeserializationSchema instead. |
Modifier and Type | Class and Description |
---|---|
class |
ChangelogCsvDeserializer
The
ChangelogCsvDeserializer contains a simple parsing logic for converting bytes into
Row of Integer and String with a RowKind . |
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<RowData> |
ChangelogCsvFormat.createRuntimeDecoder(DynamicTableSource.Context context,
DataType producedDataType) |
Modifier and Type | Method and Description |
---|---|
DecodingFormat<DeserializationSchema<RowData>> |
ChangelogCsvFormatFactory.createDecodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Constructor and Description |
---|
SocketSourceFunction(String hostname,
int port,
byte byteDelimiter,
DeserializationSchema<RowData> deserializer) |
Constructor and Description |
---|
SocketDynamicTableSource(String hostname,
int port,
byte byteDelimiter,
DecodingFormat<DeserializationSchema<RowData>> decodingFormat,
DataType producedDataType) |
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<T> |
DeserializationSchemaFactory.createDeserializationSchema(Map<String,String> properties)
Creates and configures a
DeserializationSchema using the given properties. |
Copyright © 2014–2021 The Apache Software Foundation. All rights reserved.