Modifier and Type | Method and Description |
---|---|
protected Row |
HBaseRowInputFormat.mapResultToOutType(org.apache.hadoop.hbase.client.Result res) |
Modifier and Type | Method and Description |
---|---|
DataSet<Row> |
HBaseTableSource.getDataSet(ExecutionEnvironment execEnv) |
TypeInformation<Row> |
HBaseRowInputFormat.getProducedType() |
TypeInformation<Row> |
HBaseTableSource.getReturnType() |
Modifier and Type | Method and Description |
---|---|
static TypeInformation<Row> |
Types.ROW_NAMED(String[] fieldNames,
TypeInformation<?>... types)
Returns type information for
Row with fields of the given types and
with given names. |
static TypeInformation<Row> |
Types.ROW(TypeInformation<?>... types)
Returns type information for
Row with fields of the given types. |
Modifier and Type | Method and Description |
---|---|
protected Row |
RowCsvInputFormat.fillRecord(Row reuse,
Object[] parsedValues) |
Modifier and Type | Method and Description |
---|---|
TypeInformation<Row> |
RowCsvInputFormat.getProducedType() |
Modifier and Type | Method and Description |
---|---|
protected Row |
RowCsvInputFormat.fillRecord(Row reuse,
Object[] parsedValues) |
Modifier and Type | Method and Description |
---|---|
Row |
JDBCInputFormat.nextRecord(Row row)
Stores the next resultSet row in a tuple.
|
Modifier and Type | Method and Description |
---|---|
org.apache.flink.table.sinks.TableSink<Row> |
JDBCAppendTableSink.configure(String[] fieldNames,
TypeInformation<?>[] fieldTypes) |
TypeInformation<Row> |
JDBCAppendTableSink.getOutputType() |
Modifier and Type | Method and Description |
---|---|
Row |
JDBCInputFormat.nextRecord(Row row)
Stores the next resultSet row in a tuple.
|
void |
JDBCOutputFormat.writeRecord(Row row)
Adds a record to the prepared statement.
|
Modifier and Type | Method and Description |
---|---|
void |
JDBCAppendTableSink.emitDataSet(DataSet<Row> dataSet) |
void |
JDBCAppendTableSink.emitDataStream(DataStream<Row> dataStream) |
Modifier and Type | Method and Description |
---|---|
TypeComparator<Row> |
RowTypeInfo.createComparator(int[] logicalKeyFields,
boolean[] orders,
int logicalFieldOffset,
ExecutionConfig config) |
TypeSerializer<Row> |
RowTypeInfo.createSerializer(ExecutionConfig config) |
protected CompositeType.TypeComparatorBuilder<Row> |
RowTypeInfo.createTypeComparatorBuilder() |
Modifier and Type | Method and Description |
---|---|
Row |
RowSerializer.copy(Row from) |
Row |
RowSerializer.copy(Row from,
Row reuse) |
Row |
RowSerializer.createInstance() |
Row |
RowSerializer.deserialize(DataInputView source) |
Row |
RowSerializer.deserialize(Row reuse,
DataInputView source) |
Row |
RowComparator.readWithKeyDenormalization(Row reuse,
DataInputView source) |
Modifier and Type | Method and Description |
---|---|
TypeSerializer<Row> |
RowSerializer.duplicate() |
TypeComparator<Row> |
RowComparator.duplicate() |
CompatibilityResult<Row> |
RowSerializer.ensureCompatibility(TypeSerializerConfigSnapshot<?> configSnapshot) |
Modifier and Type | Method and Description |
---|---|
int |
RowComparator.compare(Row first,
Row second) |
Row |
RowSerializer.copy(Row from) |
Row |
RowSerializer.copy(Row from,
Row reuse) |
Row |
RowSerializer.deserialize(Row reuse,
DataInputView source) |
boolean |
RowComparator.equalToReference(Row candidate) |
int |
RowComparator.hash(Row record) |
void |
RowComparator.putNormalizedKey(Row record,
MemorySegment target,
int offset,
int numBytes) |
Row |
RowComparator.readWithKeyDenormalization(Row reuse,
DataInputView source) |
void |
RowSerializer.serialize(Row record,
DataOutputView target) |
void |
RowComparator.setReference(Row toCompare) |
static void |
NullMaskUtils.writeNullMask(int len,
Row value,
DataOutputView target) |
void |
RowComparator.writeWithKeyNormalization(Row record,
DataOutputView target) |
Modifier and Type | Method and Description |
---|---|
int |
RowComparator.compareToReference(TypeComparator<Row> referencedComparator) |
Modifier and Type | Method and Description |
---|---|
protected Object[] |
CassandraRowOutputFormat.extractFields(Row record) |
Modifier and Type | Method and Description |
---|---|
Row |
AvroRowDeserializationSchema.deserialize(byte[] message) |
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
AvroRowFormatFactory.createDeserializationSchema(Map<String,String> properties) |
SerializationSchema<Row> |
AvroRowFormatFactory.createSerializationSchema(Map<String,String> properties) |
TypeInformation<Row> |
AvroRowDeserializationSchema.getProducedType() |
Modifier and Type | Method and Description |
---|---|
byte[] |
AvroRowSerializationSchema.serialize(Row row) |
Modifier and Type | Method and Description |
---|---|
static <T extends org.apache.avro.specific.SpecificRecord> |
AvroSchemaConverter.convertToTypeInfo(Class<T> avroClass)
Converts an Avro class into a nested row structure with deterministic field order and data
types that are compatible with Flink's Table & SQL API.
|
Modifier and Type | Method and Description |
---|---|
Row |
JsonRowDeserializationSchema.deserialize(byte[] message) |
Modifier and Type | Method and Description |
---|---|
DeserializationSchema<Row> |
JsonRowFormatFactory.createDeserializationSchema(Map<String,String> properties) |
SerializationSchema<Row> |
JsonRowFormatFactory.createSerializationSchema(Map<String,String> properties) |
TypeInformation<Row> |
JsonRowDeserializationSchema.getProducedType() |
Modifier and Type | Method and Description |
---|---|
boolean |
JsonRowDeserializationSchema.isEndOfStream(Row nextElement) |
byte[] |
JsonRowSerializationSchema.serialize(Row row) |
Constructor and Description |
---|
JsonRowDeserializationSchema(TypeInformation<Row> typeInfo)
Creates a JSON deserialization schema for the given type information.
|
JsonRowSerializationSchema(TypeInformation<Row> typeInfo)
Creates a JSON serialization schema for the given type information.
|
Modifier and Type | Method and Description |
---|---|
Row |
OrcRowInputFormat.nextRecord(Row reuse) |
Modifier and Type | Method and Description |
---|---|
org.apache.flink.table.sources.TableSource<Row> |
OrcTableSource.applyPredicate(List<org.apache.flink.table.expressions.Expression> predicates) |
DataSet<Row> |
OrcTableSource.getDataSet(ExecutionEnvironment execEnv) |
TypeInformation<Row> |
OrcRowInputFormat.getProducedType() |
TypeInformation<Row> |
OrcTableSource.getReturnType() |
org.apache.flink.table.sources.TableSource<Row> |
OrcTableSource.projectFields(int[] selectedFields) |
Modifier and Type | Method and Description |
---|---|
Row |
OrcRowInputFormat.nextRecord(Row reuse) |
Modifier and Type | Method and Description |
---|---|
Row |
StreamSQLTestProgram.KillMapper.map(Row value) |
Modifier and Type | Method and Description |
---|---|
DataStream<Row> |
StreamSQLTestProgram.GeneratorTableSource.getDataStream(StreamExecutionEnvironment execEnv) |
TypeInformation<Row> |
StreamSQLTestProgram.Generator.getProducedType() |
TypeInformation<Row> |
StreamSQLTestProgram.GeneratorTableSource.getReturnType() |
Modifier and Type | Method and Description |
---|---|
Row |
StreamSQLTestProgram.KillMapper.map(Row value) |
Modifier and Type | Method and Description |
---|---|
void |
StreamSQLTestProgram.Generator.run(SourceFunction.SourceContext<Row> ctx) |
Modifier and Type | Method and Description |
---|---|
protected CassandraSink<Row> |
CassandraSink.CassandraRowSinkBuilder.createSink() |
protected CassandraSink<Row> |
CassandraSink.CassandraRowSinkBuilder.createWriteAheadSink() |
TypeInformation<Row> |
CassandraAppendTableSink.getOutputType() |
Modifier and Type | Method and Description |
---|---|
protected Object[] |
CassandraRowSink.extract(Row record) |
Modifier and Type | Method and Description |
---|---|
void |
CassandraAppendTableSink.emitDataStream(DataStream<Row> dataStream) |
protected boolean |
CassandraRowWriteAheadSink.sendValues(Iterable<Row> values,
long checkpointId,
long timestamp) |
Constructor and Description |
---|
CassandraRowSinkBuilder(DataStream<Row> input,
TypeInformation<Row> typeInfo,
TypeSerializer<Row> serializer) |
CassandraRowSinkBuilder(DataStream<Row> input,
TypeInformation<Row> typeInfo,
TypeSerializer<Row> serializer) |
CassandraRowSinkBuilder(DataStream<Row> input,
TypeInformation<Row> typeInfo,
TypeSerializer<Row> serializer) |
CassandraRowWriteAheadSink(String insertQuery,
TypeSerializer<Row> serializer,
ClusterBuilder builder,
CheckpointCommitter committer) |
Modifier and Type | Method and Description |
---|---|
org.apache.flink.table.sinks.TableSink<Tuple2<Boolean,Row>> |
ElasticsearchUpsertTableSinkBase.configure(String[] fieldNames,
TypeInformation<?>[] fieldTypes) |
protected abstract SinkFunction<Tuple2<Boolean,Row>> |
ElasticsearchUpsertTableSinkBase.createSinkFunction(List<ElasticsearchUpsertTableSinkBase.Host> hosts,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.ElasticsearchUpsertSinkFunction upsertFunction) |
org.apache.flink.table.sinks.StreamTableSink<Tuple2<Boolean,Row>> |
ElasticsearchUpsertTableSinkFactoryBase.createStreamTableSink(Map<String,String> properties) |
TypeInformation<Tuple2<Boolean,Row>> |
ElasticsearchUpsertTableSinkBase.getOutputType() |
TypeInformation<Row> |
ElasticsearchUpsertTableSinkBase.getRecordType() |
Modifier and Type | Method and Description |
---|---|
protected abstract ElasticsearchUpsertTableSinkBase |
ElasticsearchUpsertTableSinkBase.copy(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
protected abstract ElasticsearchUpsertTableSinkBase |
ElasticsearchUpsertTableSinkFactoryBase.createElasticsearchUpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
void |
ElasticsearchUpsertTableSinkBase.emitDataStream(DataStream<Tuple2<Boolean,Row>> dataStream) |
void |
ElasticsearchUpsertTableSinkBase.ElasticsearchUpsertSinkFunction.process(Tuple2<Boolean,Row> element,
RuntimeContext ctx,
RequestIndexer indexer) |
Constructor and Description |
---|
ElasticsearchUpsertSinkFunction(String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory,
int[] keyFieldIndices) |
ElasticsearchUpsertTableSinkBase(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
Modifier and Type | Method and Description |
---|---|
protected SinkFunction<Tuple2<Boolean,Row>> |
Elasticsearch6UpsertTableSink.createSinkFunction(List<ElasticsearchUpsertTableSinkBase.Host> hosts,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.ElasticsearchUpsertSinkFunction upsertSinkFunction) |
Modifier and Type | Method and Description |
---|---|
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch6UpsertTableSink.copy(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions,
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory) |
protected ElasticsearchUpsertTableSinkBase |
Elasticsearch6UpsertTableSinkFactory.createElasticsearchUpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Constructor and Description |
---|
Elasticsearch6UpsertTableSink(boolean isAppendOnly,
TableSchema schema,
List<ElasticsearchUpsertTableSinkBase.Host> hosts,
String index,
String docType,
String keyDelimiter,
String keyNullLiteral,
SerializationSchema<Row> serializationSchema,
org.elasticsearch.common.xcontent.XContentType contentType,
ActionRequestFailureHandler failureHandler,
Map<ElasticsearchUpsertTableSinkBase.SinkOption,String> sinkOptions) |
Modifier and Type | Field and Description |
---|---|
protected Optional<FlinkKafkaPartitioner<Row>> |
KafkaTableSinkBase.partitioner
Partitioner to select Kafka partition for each item.
|
protected Optional<SerializationSchema<Row>> |
KafkaTableSinkBase.serializationSchema
Serialization schema for encoding records to Kafka.
|
Modifier and Type | Method and Description |
---|---|
protected FlinkKafkaConsumerBase<Row> |
Kafka08AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka08TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka08JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka011JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka011TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka011AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka010TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka010AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka010JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected abstract FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a version-specific Kafka consumer.
|
protected FlinkKafkaProducerBase<Row> |
Kafka08JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka08TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
KafkaTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
Kafka011TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka010JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka010TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka09JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka09TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected abstract SinkFunction<Row> |
KafkaTableSinkBase.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Returns the version-specific Kafka producer.
|
protected SerializationSchema<Row> |
KafkaTableSinkBase.createSerializationSchema(RowTypeInfo rowSchema)
Deprecated.
Use the constructor to pass a serialization schema instead.
|
protected SerializationSchema<Row> |
KafkaJsonTableSink.createSerializationSchema(RowTypeInfo rowSchema)
Deprecated.
|
org.apache.flink.table.sinks.StreamTableSink<Row> |
KafkaTableSourceSinkFactoryBase.createStreamTableSink(Map<String,String> properties) |
org.apache.flink.table.sources.StreamTableSource<Row> |
KafkaTableSourceSinkFactoryBase.createStreamTableSource(Map<String,String> properties) |
DataStream<Row> |
KafkaTableSourceBase.getDataStream(StreamExecutionEnvironment env)
NOTE: This method is for internal use only for defining a TableSource.
|
DeserializationSchema<Row> |
KafkaTableSourceBase.getDeserializationSchema()
Returns the deserialization schema.
|
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.getKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Returns a version-specific Kafka consumer with the start position configured.
|
TypeInformation<Row> |
KafkaTableSinkBase.getOutputType() |
TypeInformation<Row> |
KafkaTableSourceBase.getReturnType() |
Modifier and Type | Method and Description |
---|---|
protected FlinkKafkaConsumerBase<Row> |
Kafka08AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka08TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka08JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka011JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka011TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka011AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka010TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected FlinkKafkaConsumerBase<Row> |
Kafka010AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka010JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09JsonTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09AvroTableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Deprecated.
|
protected FlinkKafkaConsumerBase<Row> |
Kafka09TableSource.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema) |
protected abstract FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.createKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a version-specific Kafka consumer.
|
protected FlinkKafkaProducerBase<Row> |
Kafka08JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka08JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka08TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka08TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
KafkaTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
KafkaTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
Kafka011TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected SinkFunction<Row> |
Kafka011TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka010JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka010JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka010TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka010TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka09JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka09JsonTableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Deprecated.
|
protected FlinkKafkaProducerBase<Row> |
Kafka09TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected FlinkKafkaProducerBase<Row> |
Kafka09TableSink.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner) |
protected abstract SinkFunction<Row> |
KafkaTableSinkBase.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Returns the version-specific Kafka producer.
|
protected abstract SinkFunction<Row> |
KafkaTableSinkBase.createKafkaProducer(String topic,
Properties properties,
SerializationSchema<Row> serializationSchema,
Optional<FlinkKafkaPartitioner<Row>> partitioner)
Returns the version-specific Kafka producer.
|
protected KafkaTableSinkBase |
Kafka08TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka08TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
KafkaTableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
KafkaTableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka011TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka011TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka010TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka010TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka09TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected KafkaTableSinkBase |
Kafka09TableSourceSinkFactory.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
protected abstract KafkaTableSinkBase |
KafkaTableSourceSinkFactoryBase.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema)
Constructs the version-specific Kafka table sink.
|
protected abstract KafkaTableSinkBase |
KafkaTableSourceSinkFactoryBase.createKafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema)
Constructs the version-specific Kafka table sink.
|
protected KafkaTableSourceBase |
Kafka08TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets) |
protected KafkaTableSourceBase |
KafkaTableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets) |
protected KafkaTableSourceBase |
Kafka011TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets) |
protected KafkaTableSourceBase |
Kafka010TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets) |
protected KafkaTableSourceBase |
Kafka09TableSourceSinkFactory.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets) |
protected abstract KafkaTableSourceBase |
KafkaTableSourceSinkFactoryBase.createKafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Map<String,String> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Constructs the version-specific Kafka table source.
|
void |
KafkaTableSinkBase.emitDataStream(DataStream<Row> dataStream) |
protected FlinkKafkaConsumerBase<Row> |
KafkaTableSourceBase.getKafkaConsumer(String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Returns a version-specific Kafka consumer with the start position configured.
|
Constructor and Description |
---|
Kafka010JsonTableSink(String topic,
Properties properties,
FlinkKafkaPartitioner<Row> partitioner)
Deprecated.
Use table descriptors instead of implementation-specific classes.
|
Kafka010TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka010TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka010TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a Kafka 0.10
StreamTableSource . |
Kafka010TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.10
StreamTableSource . |
Kafka011TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka011TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka011TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a Kafka 0.11
StreamTableSource . |
Kafka011TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.11
StreamTableSource . |
Kafka08JsonTableSink(String topic,
Properties properties,
FlinkKafkaPartitioner<Row> partitioner)
Deprecated.
Use table descriptors instead of implementation-specific classes.
|
Kafka08JsonTableSink(String topic,
Properties properties,
KafkaPartitioner<Row> partitioner)
Deprecated.
This is a deprecated constructor that does not correctly handle partitioning when
producing to multiple topics. Use
Kafka08JsonTableSink.Kafka08JsonTableSink(String, Properties, FlinkKafkaPartitioner) instead. |
Kafka08TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka08TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka08TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a Kafka 0.8
StreamTableSource . |
Kafka08TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.8
StreamTableSource . |
Kafka09JsonTableSink(String topic,
Properties properties,
FlinkKafkaPartitioner<Row> partitioner)
Deprecated.
Use table descriptors instead of implementation-specific classes.
|
Kafka09JsonTableSink(String topic,
Properties properties,
KafkaPartitioner<Row> partitioner)
Deprecated.
This is a deprecated constructor that does not correctly handle partitioning when
producing to multiple topics. Use
Kafka09JsonTableSink.Kafka09JsonTableSink(String, Properties, FlinkKafkaPartitioner) instead. |
Kafka09TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka09TableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
Kafka09TableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a Kafka 0.9
StreamTableSource . |
Kafka09TableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a Kafka 0.9
StreamTableSource . |
KafkaJsonTableSink(String topic,
Properties properties,
FlinkKafkaPartitioner<Row> partitioner)
Deprecated.
Use table descriptors instead of implementation-specific classes.
|
KafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSink(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSinkBase(String topic,
Properties properties,
FlinkKafkaPartitioner<Row> partitioner)
Deprecated.
Use table descriptors instead of implementation-specific classes.
|
KafkaTableSinkBase(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSinkBase(TableSchema schema,
String topic,
Properties properties,
Optional<FlinkKafkaPartitioner<Row>> partitioner,
SerializationSchema<Row> serializationSchema) |
KafkaTableSource(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSource(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSourceBase(TableSchema schema,
Optional<String> proctimeAttribute,
List<org.apache.flink.table.sources.RowtimeAttributeDescriptor> rowtimeAttributeDescriptors,
Optional<Map<String,String>> fieldMapping,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets)
Creates a generic Kafka
StreamTableSource . |
KafkaTableSourceBase(TableSchema schema,
String topic,
Properties properties,
DeserializationSchema<Row> deserializationSchema)
Creates a generic Kafka
StreamTableSource . |
Constructor and Description |
---|
JsonRowDeserializationSchema(TypeInformation<Row> typeInfo)
Deprecated.
Please use
JsonRowDeserializationSchema in
the "flink-json" module. |
JsonRowSerializationSchema(TypeInformation<Row> typeInfo)
Deprecated.
Please use
JsonRowSerializationSchema in
the "flink-json" module. |
Modifier and Type | Method and Description |
---|---|
TypeInformation<Row> |
TableSchema.toRowType()
Converts a table schema into a (nested) type information describing a
Row . |
Modifier and Type | Method and Description |
---|---|
static String[] |
CliUtils.rowToString(Row row) |
Modifier and Type | Method and Description |
---|---|
TypedResult<List<Tuple2<Boolean,Row>>> |
Executor.retrieveResultChanges(SessionContext session,
String resultId)
Asks for the next changelog results (non-blocking).
|
List<Row> |
Executor.retrieveResultPage(String resultId,
int page)
Returns the rows that are part of the current page or throws an exception if the snapshot has been expired.
|
Modifier and Type | Method and Description |
---|---|
org.apache.flink.table.sinks.TableSink<Row> |
CollectBatchTableSink.configure(String[] fieldNames,
TypeInformation<?>[] fieldTypes) |
org.apache.flink.table.sinks.TableSink<Tuple2<Boolean,Row>> |
CollectStreamTableSink.configure(String[] fieldNames,
TypeInformation<?>[] fieldTypes) |
TypeInformation<Row> |
CollectBatchTableSink.getOutputType() |
TupleTypeInfo<Tuple2<Boolean,Row>> |
CollectStreamTableSink.getOutputType() |
TypeInformation<Row> |
CollectStreamTableSink.getRecordType() |
TypeSerializer<Row> |
CollectBatchTableSink.getSerializer()
Returns the serializer for deserializing the collected result.
|
TypedResult<List<Tuple2<Boolean,Row>>> |
LocalExecutor.retrieveResultChanges(SessionContext session,
String resultId) |
List<Row> |
LocalExecutor.retrieveResultPage(String resultId,
int page) |
Modifier and Type | Method and Description |
---|---|
void |
CollectBatchTableSink.emitDataSet(DataSet<Row> dataSet) |
void |
CollectStreamTableSink.emitDataStream(DataStream<Tuple2<Boolean,Row>> stream) |
Constructor and Description |
---|
CollectBatchTableSink(String accumulatorName,
TypeSerializer<Row> serializer) |
CollectStreamTableSink(InetAddress targetAddress,
int targetPort,
TypeSerializer<Tuple2<Boolean,Row>> serializer) |
Modifier and Type | Method and Description |
---|---|
protected List<Row> |
MaterializedCollectStreamResult.getMaterializedTable() |
TypeInformation<Row> |
CollectStreamResult.getOutputType() |
TypeInformation<Row> |
DynamicResult.getOutputType()
Returns the output type as defined by the query.
|
TypeInformation<Row> |
MaterializedCollectBatchResult.getOutputType() |
TypedResult<List<Tuple2<Boolean,Row>>> |
ChangelogResult.retrieveChanges()
Retrieves the available result records.
|
TypedResult<List<Tuple2<Boolean,Row>>> |
ChangelogCollectStreamResult.retrieveChanges() |
List<Row> |
MaterializedCollectStreamResult.retrievePage(int page) |
List<Row> |
MaterializedResult.retrievePage(int page)
Retrieves a page of a snapshotted result.
|
List<Row> |
MaterializedCollectBatchResult.retrievePage(int page) |
Modifier and Type | Method and Description |
---|---|
protected void |
MaterializedCollectStreamResult.processRecord(Tuple2<Boolean,Row> change) |
protected abstract void |
CollectStreamResult.processRecord(Tuple2<Boolean,Row> change) |
protected void |
ChangelogCollectStreamResult.processRecord(Tuple2<Boolean,Row> change) |
Constructor and Description |
---|
ChangelogCollectStreamResult(TypeInformation<Row> outputType,
ExecutionConfig config,
InetAddress gatewayAddress,
int gatewayPort) |
CollectStreamResult(TypeInformation<Row> outputType,
ExecutionConfig config,
InetAddress gatewayAddress,
int gatewayPort) |
MaterializedCollectBatchResult(TypeInformation<Row> outputType,
ExecutionConfig config) |
MaterializedCollectStreamResult(TypeInformation<Row> outputType,
ExecutionConfig config,
InetAddress gatewayAddress,
int gatewayPort,
int maxRowCount) |
MaterializedCollectStreamResult(TypeInformation<Row> outputType,
ExecutionConfig config,
InetAddress gatewayAddress,
int gatewayPort,
int maxRowCount,
int overcommitThreshold) |
Modifier and Type | Method and Description |
---|---|
Json |
Json.schema(TypeInformation<Row> schemaType)
Sets the schema using type information.
|
Modifier and Type | Method and Description |
---|---|
static Row |
Row.copy(Row row)
Creates a new Row which copied from another row.
|
static Row |
Row.of(Object... values)
Creates a new Row and assigns the given values to the Row's fields.
|
static Row |
Row.project(Row row,
int[] fields)
Creates a new Row with projected fields from another row.
|
Modifier and Type | Method and Description |
---|---|
static Row |
Row.copy(Row row)
Creates a new Row which copied from another row.
|
static Row |
Row.project(Row row,
int[] fields)
Creates a new Row with projected fields from another row.
|
Copyright © 2014–2020 The Apache Software Foundation. All rights reserved.