Modifier and Type | Class and Description |
---|---|
class |
SimpleStringSchema
Very simple serialization schema for strings.
|
class |
TypeInformationSerializationSchema<T>
A serialization and deserialization schema that uses Flink's serialization stack to transform
typed from and to byte arrays.
|
Modifier and Type | Class and Description |
---|---|
class |
AvroRowDataSerializationSchema
Serialization schema that serializes
RowData into Avro bytes. |
class |
AvroRowSerializationSchema
Serialization schema that serializes
Row into Avro bytes. |
class |
AvroSerializationSchema<T>
Serialization schema that serializes to Avro binary format.
|
class |
RegistryAvroSerializationSchema<T>
Serialization schema that serializes to Avro binary format.
|
Modifier and Type | Method and Description |
---|---|
SerializationSchema<Row> |
AvroRowFormatFactory.createSerializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
EncodingFormat<SerializationSchema<RowData>> |
AvroFormatFactory.createEncodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
ConfluentRegistryAvroSerializationSchema<T>
Serialization schema that serializes to Avro binary format that uses Confluent Schema Registry.
|
Modifier and Type | Class and Description |
---|---|
class |
CsvRowDataSerializationSchema
Serialization schema that serializes an object of Flink Table & SQL internal data structure into
a CSV bytes.
|
class |
CsvRowSerializationSchema
Serialization schema that serializes an object of Flink types into a CSV bytes.
|
Modifier and Type | Method and Description |
---|---|
SerializationSchema<Row> |
CsvRowFormatFactory.createSerializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
EncodingFormat<SerializationSchema<RowData>> |
CsvFormatFactory.createEncodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Class and Description |
---|---|
class |
JsonRowDataSerializationSchema
Serialization schema that serializes an object of Flink internal data structure into a JSON
bytes.
|
class |
JsonRowSerializationSchema
Serialization schema that serializes an object of Flink types into a JSON bytes.
|
Modifier and Type | Method and Description |
---|---|
SerializationSchema<Row> |
JsonRowFormatFactory.createSerializationSchema(Map<String,String> properties) |
Modifier and Type | Method and Description |
---|---|
EncodingFormat<SerializationSchema<RowData>> |
JsonFormatFactory.createEncodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Method and Description |
---|---|
EncodingFormat<SerializationSchema<RowData>> |
CanalJsonFormatFactory.createEncodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Method and Description |
---|---|
EncodingFormat<SerializationSchema<RowData>> |
DebeziumJsonFormatFactory.createEncodingFormat(DynamicTableFactory.Context context,
ReadableConfig formatOptions) |
Modifier and Type | Method and Description |
---|---|
DataStreamSink<T> |
DataStream.writeToSocket(String hostName,
int port,
SerializationSchema<T> schema)
Writes the DataStream to a socket as a byte array.
|
Constructor and Description |
---|
SocketClientSink(String hostName,
int port,
SerializationSchema<IN> schema)
Creates a new SocketClientSink.
|
SocketClientSink(String hostName,
int port,
SerializationSchema<IN> schema,
int maxNumRetries)
Creates a new SocketClientSink that retries connections upon failure up to a given number of
times.
|
SocketClientSink(String hostName,
int port,
SerializationSchema<IN> schema,
int maxNumRetries,
boolean autoflush)
Creates a new SocketClientSink that retries connections upon failure up to a given number of
times.
|
Modifier and Type | Method and Description |
---|---|
protected abstract ElasticsearchUpsertTableSinkBase |
ElasticsearchUpsertTableSinkBase.copy(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
protected abstract ElasticsearchUpsertTableSinkBase |
ElasticsearchUpsertTableSinkFactoryBase.createElasticsearchUpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Constructor and Description |
---|
ElasticsearchUpsertSinkFunction(IndexGenerator indexGenerator,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory,
int[] keyFieldIndices) |
ElasticsearchUpsertTableSinkBase(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
Modifier and Type | Method and Description |
---|---|
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch6UpsertTableSink.copy(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch6UpsertTableSinkFactory.createElasticsearchUpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Constructor and Description |
---|
Elasticsearch6UpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Modifier and Type | Method and Description |
---|---|
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch7UpsertTableSink.copy(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch7UpsertTableSinkFactory.createElasticsearchUpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Constructor and Description |
---|
Elasticsearch7UpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Modifier and Type | Method and Description |
---|---|
<IN> PubSubSink.ProjectNameBuilder<IN> |
PubSubSink.SerializationSchemaBuilder.withSerializationSchema(SerializationSchema<IN> deserializationSchema)
Set the SerializationSchema used to Serialize objects to be added as payloads of
PubSubMessages.
|
Modifier and Type | Field and Description |
---|---|
protected SerializationSchema<Row> |
KafkaTableSinkBase.serializationSchema
Serialization schema for encoding records to Kafka.
|
Modifier and Type | Method and Description |
---|---|
protected SinkFunction<Row> |
KafkaTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
Kafka011TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka010TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected abstract SinkFunction<Row> |
KafkaTableSinkBase.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Returns the version-specific Kafka producer.
|
protected KafkaTableSinkBase |
KafkaTableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka011TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka010TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected abstract KafkaTableSinkBase |
KafkaTableSourceSinkFactoryBase.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema)
Constructs the version-specific Kafka table sink.
|
static <T> FlinkKafkaProducer010.FlinkKafkaProducer010Configuration<T> |
FlinkKafkaProducer010.writeToKafkaWithTimestamps(DataStream<T> inStream,
String topicId,
SerializationSchema<T> serializationSchema,
Properties producerConfig)
|
Constructor and Description |
---|
FlinkKafkaProducer(String topicId,
SerializationSchema<IN> serializationSchema,
Properties producerConfig)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer(String topicId,
SerializationSchema<IN> serializationSchema,
Properties producerConfig,
FlinkKafkaPartitioner<IN> customPartitioner,
FlinkKafkaProducer.Semantic semantic,
int kafkaProducersPoolSize)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer(String topicId,
SerializationSchema<IN> serializationSchema,
Properties producerConfig,
Optional<FlinkKafkaPartitioner<IN>> customPartitioner)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer(String brokerList,
String topicId,
SerializationSchema<IN> serializationSchema)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer010(String topicId,
SerializationSchema<T> serializationSchema,
Properties producerConfig)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer010(String topicId,
SerializationSchema<T> serializationSchema,
Properties producerConfig,
FlinkKafkaPartitioner<T> customPartitioner)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer010(String brokerList,
String topicId,
SerializationSchema<T> serializationSchema)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer011(String topicId,
SerializationSchema<IN> serializationSchema,
Properties producerConfig)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer011(String topicId,
SerializationSchema<IN> serializationSchema,
Properties producerConfig,
Optional<FlinkKafkaPartitioner<IN>> customPartitioner)
Creates a FlinkKafkaProducer for a given topic.
|
FlinkKafkaProducer011(String brokerList,
String topicId,
SerializationSchema<IN> serializationSchema)
Creates a FlinkKafkaProducer for a given topic.
|
Kafka010TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka011TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSinkBase(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Modifier and Type | Method and Description |
---|---|
SerializationSchema<T> |
KeyedSerializationSchemaWrapper.getSerializationSchema() |
Constructor and Description |
---|
KafkaSerializationSchemaWrapper(String topic,
FlinkKafkaPartitioner<T> partitioner,
boolean writeTimestamp,
SerializationSchema<T> serializationSchema) |
KeyedSerializationSchemaWrapper(SerializationSchema<T> serializationSchema) |
Modifier and Type | Field and Description |
---|---|
protected EncodingFormat<SerializationSchema<RowData>> |
KafkaDynamicSinkBase.encodingFormat
Sink format for encoding records to Kafka.
|
Constructor and Description |
---|
Kafka010DynamicSink(DataType consumedDataType,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<RowData>> partitioner,
EncodingFormat<SerializationSchema<RowData>> encodingFormat) |
Kafka011DynamicSink(DataType consumedDataType,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<RowData>> partitioner,
EncodingFormat<SerializationSchema<RowData>> encodingFormat) |
KafkaDynamicSink(DataType consumedDataType,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<RowData>> partitioner,
EncodingFormat<SerializationSchema<RowData>> encodingFormat) |
KafkaDynamicSinkBase(DataType consumedDataType,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<RowData>> partitioner,
EncodingFormat<SerializationSchema<RowData>> encodingFormat) |
Constructor and Description |
---|
FlinkKinesisProducer(SerializationSchema<OUT> schema,
Properties configProps)
Create a new FlinkKinesisProducer.
|
Modifier and Type | Field and Description |
---|---|
protected SerializationSchema<IN> |
RMQSink.schema |
Constructor and Description |
---|
RMQSink(RMQConnectionConfig rmqConnectionConfig,
SerializationSchema<IN> schema,
RMQSinkPublishOptions<IN> publishOptions) |
RMQSink(RMQConnectionConfig rmqConnectionConfig,
SerializationSchema<IN> schema,
RMQSinkPublishOptions<IN> publishOptions,
SerializableReturnListener returnListener) |
RMQSink(RMQConnectionConfig rmqConnectionConfig,
String queueName,
SerializationSchema<IN> schema) |
Modifier and Type | Class and Description |
---|---|
class |
EventDeSerializer
A serializer and deserializer for the
Event type. |
Modifier and Type | Class and Description |
---|---|
class |
KafkaEventSchema
The serialization schema for the
KafkaEvent type. |
Modifier and Type | Interface and Description |
---|---|
interface |
SerializationSchema<T>
Deprecated.
Use
SerializationSchema instead. |
Modifier and Type | Method and Description |
---|---|
SerializationSchema<T> |
SerializationSchemaFactory.createSerializationSchema(Map<String,String> properties)
Creates and configures a [[SerializationSchema]] using the given properties.
|
Copyright © 2014–2021 The Apache Software Foundation. All rights reserved.