Class and Description |
---|
org.apache.flink.streaming.util.serialization.AbstractDeserializationSchema
Use
AbstractDeserializationSchema instead. |
org.apache.flink.streaming.api.functions.AscendingTimestampExtractor
Extend
AscendingTimestampExtractor instead. |
org.apache.flink.api.java.typeutils.runtime.AvroSerializer |
org.apache.flink.formats.avro.typeutils.AvroSerializer.AvroSerializerConfigSnapshot
The
AvroSerializer.AvroSchemaSerializerConfigSnapshot should be used instead. |
org.apache.flink.api.java.typeutils.AvroTypeInfo
Please use
org.apache.flink.formats.avro.typeutils.AvroTypeInfo
in the flink-avro module. This class will be removed in the near future. |
org.apache.flink.streaming.runtime.io.BufferSpiller |
org.apache.flink.streaming.runtime.io.BufferSpiller.SpilledBufferOrEventSequence |
org.apache.flink.batch.connectors.cassandra.CassandraOutputFormat
Please use CassandraTupleOutputFormat instead.
|
org.apache.flink.streaming.connectors.fs.DateTimeBucketer
use
DateTimeBucketer instead. |
org.apache.flink.streaming.runtime.operators.ExtractTimestampsOperator |
org.apache.flink.streaming.api.functions.source.FileMonitoringFunction
Internal class deprecated in favour of
ContinuousFileMonitoringFunction . |
org.apache.flink.streaming.api.functions.source.FileReadFunction
Internal class deprecated in favour of
ContinuousFileMonitoringFunction . |
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer081 |
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer082 |
org.apache.flink.streaming.connectors.kafka.partitioner.FlinkKafkaDelegatePartitioner
Delegate for
KafkaPartitioner , use FlinkKafkaPartitioner instead |
org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer
Use
FlinkKafkaProducer08 . |
org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer010.FlinkKafkaProducer010Configuration
This class is deprecated since the factory methods
writeToKafkaWithTimestamps
for the producer are also deprecated. |
org.apache.flink.streaming.api.functions.windowing.FoldApplyAllWindowFunction
will be removed in a future version
|
org.apache.flink.streaming.api.functions.windowing.FoldApplyProcessAllWindowFunction
will be removed in a future version
|
org.apache.flink.streaming.api.functions.windowing.FoldApplyProcessWindowFunction
will be removed in a future version
|
org.apache.flink.streaming.api.functions.windowing.FoldApplyWindowFunction
will be removed in a future version
|
org.apache.flink.api.common.state.FoldingStateDescriptor
will be removed in a future version in favor of
AggregatingStateDescriptor |
org.apache.flink.queryablestate.client.state.ImmutableFoldingState |
org.apache.flink.runtime.rest.handler.job.metrics.JobVertexMetricsHandler
This class is subsumed by
SubtaskMetricsHandler and is only kept for
backwards-compatibility. |
org.apache.flink.runtime.rest.handler.legacy.metrics.JobVertexMetricsHandler
This class is subsumed by
SubtaskMetricsHandler and is only kept for backwards-compatibility. |
org.apache.flink.streaming.util.serialization.JSONDeserializationSchema
Please use
JsonNodeDeserializationSchema in the "flink-json" module. |
org.apache.flink.streaming.util.serialization.JsonRowDeserializationSchema
Please use
JsonRowDeserializationSchema in
the "flink-json" module. |
org.apache.flink.streaming.util.serialization.JsonRowSerializationSchema
Please use
JsonRowSerializationSchema in
the "flink-json" module. |
org.apache.flink.streaming.connectors.kafka.Kafka010AvroTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka010AvroTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka010JsonTableSink
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka010JsonTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka010JsonTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka011AvroTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka011AvroTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka011JsonTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka011JsonTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka08AvroTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka08AvroTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka08JsonTableSink
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka08JsonTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka08JsonTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka09AvroTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka09AvroTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka09JsonTableSink
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka09JsonTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.Kafka09JsonTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.KafkaAvroTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.KafkaAvroTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.KafkaJsonTableSink
Use table descriptors instead of implementation-specific classes.
|
org.apache.flink.streaming.connectors.kafka.KafkaJsonTableSource
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.KafkaJsonTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.connectors.kafka.partitioner.KafkaPartitioner
This partitioner does not handle partitioning properly in the case of
multiple topics, and has been deprecated. Please use
FlinkKafkaPartitioner instead. |
org.apache.flink.streaming.connectors.kafka.KafkaTableSource.Builder
Use the
Kafka descriptor together
with descriptors for schema and format instead. Descriptors allow for
implementation-agnostic definition of tables. See also
TableEnvironment.connect(ConnectorDescriptor) . |
org.apache.flink.streaming.api.operators.LegacyKeyedProcessOperator
Replaced by
KeyedProcessOperator which takes KeyedProcessFunction |
org.apache.flink.cep.nfa.NFA.NFASerializer |
org.apache.flink.cep.nfa.NFA.NFASerializerConfigSnapshot |
org.apache.flink.streaming.connectors.fs.NonRollingBucketer
use
BasePathBucketer instead. |
org.apache.flink.streaming.api.functions.sink.OutputFormatSinkFunction
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.api.common.functions.RichFoldFunction
use
RichAggregateFunction instead |
org.apache.flink.streaming.api.functions.windowing.RichProcessAllWindowFunction
use
ProcessAllWindowFunction instead |
org.apache.flink.streaming.api.functions.windowing.RichProcessWindowFunction
use
ProcessWindowFunction instead |
org.apache.flink.streaming.connectors.fs.RollingSink
use
BucketingSink instead. |
org.apache.flink.cep.nfa.SharedBuffer
everything in this class is deprecated. Those are only migration procedures from older versions.
|
org.apache.flink.streaming.util.serialization.SimpleStringSchema
Use
SimpleStringSchema instead. |
org.apache.flink.streaming.api.windowing.assigners.SlidingTimeWindows
Please use
SlidingEventTimeWindows . |
org.apache.flink.streaming.api.datastream.SplitStream |
org.apache.flink.streaming.api.operators.StreamGroupedFold
will be removed in a future version
|
org.apache.flink.runtime.checkpoint.TaskState
Internal class for savepoint backwards compatibility. Don't use for other purposes.
|
org.apache.flink.streaming.api.windowing.assigners.TumblingTimeWindows
Please use
TumblingEventTimeWindows . |
org.apache.flink.streaming.util.serialization.TypeInformationSerializationSchema
Use
TypeInformationSerializationSchema instead. |
org.apache.flink.streaming.api.functions.sink.WriteFormat
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.streaming.api.functions.sink.WriteFormatAsCsv
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.streaming.api.functions.sink.WriteFormatAsText
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.streaming.api.functions.sink.WriteSinkFunction
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.streaming.api.functions.sink.WriteSinkFunctionByMillis
Please use the
BucketingSink for writing to files from a streaming program. |
org.apache.flink.runtime.rest.messages.YarnCancelJobTerminationHeaders
This should be removed once we can send arbitrary REST calls via the Yarn proxy.
|
org.apache.flink.runtime.rest.messages.YarnStopJobTerminationHeaders
This should be removed once we can send arbitrary REST calls via the Yarn proxy.
|
Enum and Description |
---|
org.apache.flink.test.util.TestBaseUtils.CodebaseType
Will be irrelevant once the legacy mode has been removed.
|
Enum Constant and Description |
---|
org.apache.flink.api.common.state.StateDescriptor.Type.UNKNOWN
Enum for migrating from old checkpoints/savepoint versions.
|
Copyright © 2014–2020 The Apache Software Foundation. All rights reserved.