Collator collator
String prefix
ExecutionMode executionMode
ExecutionConfig.ClosureCleanerLevel closureCleanerLevel
int parallelism
int maxParallelism
int numberOfExecutionRetries
boolean forceKryo
boolean disableGenericTypes
boolean enableAutoGeneratedUids
boolean objectReuse
boolean autoTypeRegistrationEnabled
boolean forceAvro
long autoWatermarkInterval
long latencyTrackingInterval
boolean isLatencyTrackingConfigured
long executionRetryDelay
RestartStrategies.RestartStrategyConfiguration restartStrategyConfiguration
long taskCancellationIntervalMillis
long taskCancellationTimeoutMillis
boolean useSnapshotCompression
InputDependencyConstraint defaultInputDependencyConstraint
ExecutionConfig.GlobalJobParameters globalJobParameters
LinkedHashMap<K,V> registeredTypesWithKryoSerializers
LinkedHashMap<K,V> registeredTypesWithKryoSerializerClasses
LinkedHashMap<K,V> defaultKryoSerializers
LinkedHashMap<K,V> defaultKryoSerializerClasses
LinkedHashSet<E> registeredKryoTypes
LinkedHashSet<E> registeredPojoTypes
com.esotericsoftware.kryo.Serializer<T> serializer
long count
double sum
double localValue
double max
double min
int localValue
int max
int min
long localValue
long max
long min
DoubleValue wrapper
double sum
long sum
TypeComparator<T> typeComparator
Object[][] boundaries
TypeComparator<T>[] flatComparators
Object[] keys
SerializableTimestampAssigner<T> assigner
long timestamp
boolean target
String sinkIdentifier
long blockSize
long readRecords
long blockSize
Path filePath
FileInputFormat.supportsMultiPaths()
and use FileInputFormat.getFilePaths()
and FileInputFormat.setFilePaths(Path...)
.Path[] filePaths
long minSplitSize
int numSplits
long openTimeout
boolean unsplittable
boolean enumerateNestedFiles
FilePathFilter filesFilter
Path outputFilePath
FileSystem.WriteMode writeMode
FileOutputFormat.OutputDirectoryMode outputDirectoryMode
boolean lineDelimiterIsLinebreak
Class<T>[] fieldTypes
boolean[] fieldIncluded
byte[] fieldDelim
String fieldDelimString
boolean lenient
boolean skipFirstLineAsHeader
boolean quotedStringParsing
byte quoteCharacter
byte[] commentPrefix
String commentPrefixString
int partitionNumber
InputFormat<OT,T extends InputSplit> replicatedIF
long globalRateBytesPerSecond
long localRateBytesPerSecond
RuntimeContext runtimeContext
org.apache.flink.shaded.guava18.com.google.common.util.concurrent.RateLimiter rateLimiter
Map<K,V> fieldMapping1
Map<K,V> fieldMapping2
FieldSet readFields1
FieldSet readFields2
private Object readResolve()
Resource cpuCores
MemorySize taskHeapMemory
MemorySize taskOffHeapMemory
MemorySize managedMemory
Map<K,V> extendedResources
long count
BulkIterationBase.TerminationCriterionAggregator aggregator
Object userCodeObject
String name
BigDecimal value
int restartAttempts
Time delayBetweenAttemptsInterval
TypeInformation<T> type
DeserializationSchema
.String charsetName
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeInformation<T> typeInfo
TypeInformationSerializationSchema.getProducedType()
.TypeSerializer<T> serializer
AggregateFunction<IN,ACC,OUT> aggFunction
ReduceFunction<T> reduceFunction
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
String name
AtomicReference<V> serializerAtomicReference
StateDescriptor.initializeSerializerUnlessSet(ExecutionConfig)
method is called.TypeInformation<T> typeInfo
String queryableStateName
StateTtlConfig ttlConfig
StateTtlConfig.UpdateType updateType
StateTtlConfig.StateVisibility stateVisibility
StateTtlConfig.TtlTimeCharacteristic ttlTimeCharacteristic
Time ttl
StateTtlConfig.CleanupStrategies cleanupStrategies
int cleanupSize
boolean runCleanupForEveryRecord
long queryTimeAfterNumEntries
TimeUnit unit
long size
Class<T> arrayClass
TypeInformation<T> componentInfo
TypeSerializer<T>[] fieldSerializers
CompositeSerializer.PrecomputedParameters precomputed
boolean immutableTargetType
boolean immutable
int length
boolean stateful
TypeComparator<T> comparator1
TypeComparator<T> comparator2
TypeComparator<T>[] comparators1
TypeComparator<T>[] comparators2
Object[] referenceKeyFields
Object[] candidateKeyFields
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
org.apache.flink.api.common.typeutils.SingleThreadAccessCheckingTypeSerializer.SingleThreadAccessChecker singleThreadAccessChecker
TypeSerializer<T> originalSerializer
byte[] actualBytes
Throwable originalError
boolean ascendingComparison
TypeComparator<T>[] comparators
boolean ascendingComparison
BooleanValue reference
BooleanValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
ByteValue reference
ByteValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
CharValue reference
CharValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
DoubleValue reference
DoubleValue tempReference
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> enumClass
Map<K,V> valueToOrdinal
Enum.ordinal()
. This allows us to maintain backwards compatibility for previous serialized
data in the case that the order of enum constants was changed or new constants were added.
On a fresh start with no reconfiguration, the ordinals would simply be identical to the enum constants actual ordinals. Ordinals may change after reconfiguration.
Enum<E extends Enum<E>>[] values
EnumSerializer.valueToOrdinal
map. Serves as a bidirectional map to have fast access from ordinal to
value. May be reordered after reconfiguration.boolean ascendingComparison
FloatValue reference
FloatValue tempReference
TypeComparator<T>[] comparators
Class<T> componentClass
TypeSerializer<T> componentSerializer
boolean ascendingComparison
IntValue reference
IntValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> elementSerializer
boolean ascendingComparison
LocalDateComparator[] comparators
boolean ascendingComparison
LocalDateComparator dateComparator
LocalTimeComparator timeComparator
LocalDateTimeComparator[] comparators
boolean ascendingComparison
LongValue reference
LongValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeComparator<T>[] comparators
boolean ascendingComparison
ShortValue reference
ShortValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
StringValue reference
StringValue tempReference
TypeComparator<T>[] comparators
TypeComparator<T>[] comparators
boolean ascending
BasicTypeComparator<T extends Comparable<T>> comparator
int subtaskId
String location
long from
long to
long count
long checksum
String id
long counter
long checksum
String id
TypeSerializer<T> serializer
SerializedListAccumulator<T> accumulator
String id
long counter
int count
TextOutputFormat.TextFormatter<IN> formatter
boolean withReplacement
int numSample
long seed
boolean withReplacement
int numSample
long seed
boolean withReplacement
double fraction
long seed
int[] fields
int[] fields
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapred.JobConf jobConf
org.apache.hadoop.mapred.OutputFormat<K,V> mapredOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapreduce.InputFormat<K,V> mapreduceInputFormat
Class<T> keyClass
Class<T> valueClass
Configuration configuration
boolean fetched
boolean hasNext
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration configuration
org.apache.hadoop.mapreduce.OutputFormat<K,V> mapreduceOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
AbstractID intermediateDataSetId
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeSerializer<T> serializer
TypeSerializer<T> typeSerializer
int id
SplittableIterator<T> source
PrintSinkOutputWriter<IN> writer
int arity
TypeInformation<T>[] fieldTypeInfos
int[] fieldPosMap
boolean emptyColumnAsNull
String partitionMarker
String charsetName
String charsetName
String charsetName
boolean skipInvalidLines
TupleSerializerBase<T> tupleSerializer
TypeSerializer<T> serializer
TypeSerializer<T> serializer
int[] fields
boolean[] isFromFirst
Tuple outTuple
CombineFunction<IN,OUT> wrappedFunction
RichGroupReduceFunction<IN,OUT> wrappedFunction
KeySelector<IN,KEY> keySelector1
KeySelector<IN,KEY> keySelector2
Tuple3<T0,T1,T2> tuple
Function wrappedFunction
long trueCount
long falseCount
long nullCount
double value
double delta
double max
double min
CompensatedSum sum
float max
float min
CompensatedSum sum
int max
int min
int sum
long max
long min
long nonMissingCount
long nullCount
long nanCount
long infinityCount
Aggregator<T,R> min
Aggregator<T,R> max
Aggregator<T,R> sum
CompensatedSum mean
CompensatedSum m2
The algorithm is described in: "Scalable and Numerically Stable Descriptive Statistics in SystemML", Tian et al, International Conference on Data Engineering 2012
long nonNullCount
long nullCount
short max
short min
short sum
long nonNullCount
long nullCount
long emptyCount
int minStringLength
int maxStringLength
CompensatedSum meanLength
Aggregator<T,R>[] columnAggregators
Aggregator<T,R> aggregator
private Object readResolve() throws ObjectStreamException
ObjectStreamException
Object f0
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f24
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
TypeInformation<T> leftType
TypeInformation<T> rightType
TypeInformation<T> elementTypeInfo
TypeInformation<T> keyTypeInfo
TypeInformation<T> valueTypeInfo
String functionName
InvalidTypesException typeException
Class<T> arrayType
TypeInformation<T> componentInfo
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeInformation<T> type
PojoField[] fields
int totalFields
String[] fieldNames
boolean[] comparatorOrders
String[] fieldNames
TypeInformation<T>[] types
int totalFields
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
boolean ascending
Class<T> type
TypeSerializer<T> serializer
TypeComparator<T>[] comparators
Class<T> registeredClass
This can be a dummy class KryoRegistrationSerializerConfigSnapshot.DummyRegisteredClass
if the class no longer exists
when this registration instance was restored.
Class<T> serializerClass
KryoRegistration.SerializerDefinitionType.CLASS
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if the serializer class no
longer exists when this registration instance was restored.
ExecutionConfig.SerializableSerializer<T extends com.esotericsoftware.kryo.Serializer<?> & Serializable> serializableSerializerInstance
KryoRegistration.SerializerDefinitionType.INSTANCE
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if the serializer class no
longer exists or is no longer valid when this registration instance was restored.
KryoRegistration.SerializerDefinitionType serializerDefinitionType
TypeSerializer<T> originalSerializer
byte[] padding
TypeComparator<T> wrappedComparator
boolean order
int flatFields
boolean nullReference
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T> serializer
Class<T> type
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> clazz
TypeSerializer<T>[] fieldSerializers
int numFields
LinkedHashMap<K,V> registeredClasses
TypeSerializer<T>[] registeredSerializers
ExecutionConfig executionConfig
int arity
int[] keyPositions
NullAwareComparator<T>[] comparators
TypeSerializer<T>[] serializers
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
boolean[] mask1
boolean[] mask2
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
boolean legacyModeEnabled
int legacyOffset
TypeSerializer<T>[] fieldSerializers
int arity
TypeComparator<T> comparator
TypeSerializer<T> serializer
boolean firstSerializer
Class<T> clazz
int[] keyPositions
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T>[] serializers
Class<T> tupleClass
TypeSerializer<T>[] fieldSerializers
int arity
int length
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Currently, we only have one single registration for the value type. Nevertheless, we keep this information here for future compatibility.
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
LinkedHashMap<K,V> defaultSerializers
LinkedHashMap<K,V> defaultSerializerClasses
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Class<T> type
LinkedHashMap<K,V> registeredTypesWithSerializers
LinkedHashMap<K,V> registeredTypesWithSerializerClasses
LinkedHashSet<E> registeredTypes
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Object[] extractedKeys
Class<T> clazz
TypeInformation<T>[] typeParamTypeInfos
scala.collection.Seq<A> org$apache$flink$api$scala$typeutils$CaseClassTypeInfo$$fieldTypes
scala.collection.Seq<A> fieldNames
String REGEX_INT_FIELD
String REGEX_STR_FIELD
String REGEX_FIELD
String REGEX_NESTED_FIELDS
String REGEX_NESTED_FIELDS_WILDCARD
Pattern PATTERN_NESTED_FIELDS
Pattern PATTERN_NESTED_FIELDS_WILDCARD
Pattern PATTERN_INT_FIELD
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
Class<T> clazz
TypeInformation<T> leftTypeInfo
TypeInformation<T> rightTypeInfo
boolean ascComp
IntComparator intComparator
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
scala.Enumeration enum
private Object readResolve()
TypeSerializer<T> elemSerializer
private Object readResolve()
boolean ascending
TypeComparator<T> typeComparator
scala.Option<A> reference
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
private Object readResolve()
byte ZeroInByte
byte OneInByte
TypeInformation<T> elemTypeInfo
private void readObject(ObjectInputStream arg0)
private Object readResolve()
private void readObject(ObjectInputStream arg0)
TypeSerializer<T> elementSerializer
String cbfCode
int classLoaderHash
scala.ref.WeakReference<T> classLoaderRef
String cbfCode
private Object readResolve()
ClassLoader classLoader
String code
private Object readResolve()
private Object readResolve()
org.apache.flink.shaded.guava18.com.google.common.cache.Cache<K,V> CACHE
Class<T> clazz
TypeInformation<T> elementTypeInfo
TypeSerializer<T> elemSerializer
TypeSerializer<T> throwableSerializer
private Object readResolve()
TypeInformation<T> elemTypeInfo
private Object readResolve()
org.slf4j.Logger logger
String query
ClusterBuilder builder
String insertQuery
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> inputClass
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> outputClass
String latchFilePath
long numRecords
long numKeys
long recordsPerPartition
long keysPerPartition
long recordCnt
int partitionId
boolean infinite
PatternFlatSelectFunction<IN,OUT> flatSelectFunction
PatternSelectFunction<IN,OUT> selectFunction
PatternFlatTimeoutFunction<IN,OUT> flatTimeoutFunction
OutputTag<T> timedOutPartialMatchesTag
PatternTimeoutFunction<IN,OUT> timeoutFunction
OutputTag<T> timedOutPartialMatchesTag
int[] deweyNumber
TypeSerializer<T> sharedBufferSerializer
TypeSerializer<T> eventSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> versionSerializer
NFAStateSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> nodeIdSerializer
TypeSerializer<T> eventIdSerializer
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeSerializer<T> versionSerializer
String name
State.StateType stateType
Collection<E> stateTransitions
StateTransitionAction action
State<T> sourceState
State<T> targetState
IterativeCondition<T> condition
TypeSerializer<T> elementSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> eventIdSerializer
NodeId.NodeIdSerializer.readObject(ObjectInputStream)
.private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> nodeIdSerializer
SharedBufferEdge.SharedBufferEdgeSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> deweyNumberSerializer
ListSerializer<T> edgesSerializer
boolean isProcessingTime
TypeSerializer<T> inputSerializer
NFACompiler.NFAFactory<T> nfaFactory
EventComparator<T> comparator
OutputTag<T> lateDataOutputTag
OutputTag
to use for late arriving events. Elements with timestamp smaller than the
current watermark will be emitted to this.AfterMatchSkipStrategy afterMatchSkipStrategy
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T> original
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T>[] nestedConditions
ApplicationStatus status
Configuration backingConfig
String prefix
long bytes
org.apache.flink.connector.file.sink.FileSink.BucketsBuilder<IN,T extends org.apache.flink.connector.file.sink.FileSink.BucketsBuilder<IN,T>> bucketsBuilder
Path basePath
long bucketCheckInterval
BulkWriter.Factory<T> writerFactory
FileWriterBucketFactory<IN> bucketFactory
BucketAssigner<IN,BucketID> bucketAssigner
CheckpointRollingPolicy<IN,BucketID> rollingPolicy
OutputFileConfig outputFileConfig
Path basePath
long bucketCheckInterval
Encoder<IN> encoder
FileWriterBucketFactory<IN> bucketFactory
BucketAssigner<IN,BucketID> bucketAssigner
RollingPolicy<IN,BucketID> rollingPolicy
OutputFileConfig outputFileConfig
InProgressFileWriter.PendingFileRecoverable pendingFile
InProgressFileWriter.InProgressFileRecoverable inProgressFileToCleanup
Path[] inputPaths
FileEnumerator.Provider enumeratorFactory
FileSplitAssigner.Provider assignerFactory
BulkFormat<T,SplitT extends FileSourceSplit> readerFormat
ContinuousEnumerationSettings continuousEnumerationSettings
java.time.Duration discoveryInterval
String id
Path filePath
long offset
long length
String[] hostnames
CheckpointedPosition readerPosition
FileRecordFormat<T> fileFormat
StreamFormat<T> streamFormat
String charsetName
long offset
long recordsAfterOffset
long bufferFlushMaxSizeInBytes
long bufferFlushMaxRows
long bufferFlushIntervalMillis
Integer parallelism
String hTableName
byte[] serializedConfig
long bufferFlushMaxSizeInBytes
long bufferFlushMaxMutations
long bufferFlushIntervalMillis
HBaseMutationConverter<T> mutationConverter
AtomicReference<V> failureThrowable
BufferedMutator.ExceptionListener
if a Throwable
was thrown.
Errors will be checked and rethrown before processing each input element, and when the sink is closed.
HBaseTableSchema schema
HBaseTableSchema schema
String nullStringLiteral
String hTableName
byte[] serializedConfig
HBaseTableSchema hbaseTableSchema
String hTableName
byte[] serializedConfig
HBaseTableSchema hbaseTableSchema
String nullStringLiteral
byte[] tableName
byte[] startRow
byte[] endRow
boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] currentRow
long scannedRows
byte[] serializedConfig
String tableName
HBaseTableSchema schema
String nullStringLiteral
String tableName
HBaseTableSchema schema
boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] currentRow
long scannedRows
byte[] serializedConfig
String tableName
HBaseTableSchema schema
String nullStringLiteral
String tableName
HBaseTableSchema schema
long batchIntervalMs
int batchSize
int maxRetries
String username
String password
String drivername
String dbURL
String queryTemplate
int resultSetType
int resultSetConcurrency
RowTypeInfo rowTypeInfo
int fetchSize
Boolean autoCommit
boolean hasNext
Object[][] parameterValues
JdbcConnectionProvider connectionProvider
AbstractJdbcOutputFormat<T> outputFormat
JdbcExecutionOptions executionOptions
JdbcBatchingOutputFormat.StatementExecutorFactory<T extends JdbcBatchStatementExecutor<?>> statementExecutorFactory
JdbcBatchingOutputFormat.RecordExtractor<F,T> jdbcRecordExtractor
JdbcConnectionOptions jdbcOptions
RowType rowType
org.apache.flink.connector.jdbc.internal.converter.AbstractJdbcRowConverter.JdbcDeserializationConverter[] toInternalConverters
org.apache.flink.connector.jdbc.internal.converter.AbstractJdbcRowConverter.JdbcSerializationConverter[] toExternalConverters
LogicalType[] fieldTypes
String[] fieldNames
String[] keyFields
String tableName
JdbcDialect dialect
String query
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
String tableName
JdbcDialect dialect
JdbcOptions jdbcOptions
JdbcExecutionOptions executionOptions
JdbcDmlOptions dmlOptions
TypeInformation<T> rowDataTypeInformation
DataType[] fieldDataTypes
String query
String drivername
String dbURL
String username
String password
TypeInformation<T>[] keyTypes
int[] keySqlTypes
String[] fieldNames
String[] keyNames
TypeInformation<T>[] fieldTypes
int[] outputSqlTypes
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
JdbcConnectionOptions connectionOptions
int fetchSize
Boolean autoCommit
Object[][] parameterValues
String queryTemplate
int resultSetType
int resultSetConcurrency
JdbcRowConverter rowConverter
TypeInformation<T> rowDataTypeInfo
String query
String drivername
String dbURL
String username
String password
DataType[] keyTypes
String[] keyNames
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
JdbcDialect jdbcDialect
JdbcRowConverter jdbcRowConverter
JdbcRowConverter lookupKeyRowConverter
KafkaSubscriber subscriber
OffsetsInitializer startingOffsetsInitializer
OffsetsInitializer stoppingOffsetsInitializer
Boundedness boundedness
KafkaRecordDeserializer<T> deserializationSchema
Properties props
JobConfWrapper jobConfWrapper
DataFormatConverters.DataFormatConverter<Internal,External>[] partitionConverters
HiveObjectConversion[] hiveObjectConversions
HiveObjectConversion[] partColConversions
JobConfWrapper jobConfWrapper
List<E> partitionKeys
ContinuousPartitionFetcher<P,T extends Comparable<T>> fetcher
HiveTableSource.HiveContinuousPartitionFetcherContext<T extends Comparable<T>> fetcherContext
ObjectPath tablePath
List<E> partitions
JobConfWrapper jobConfWrapper
JobConfWrapper conf
String hiveVersion
String database
String tableName
CachedSerializedValue<T> storageDescriptor
Map<K,V> partitionSpec
Properties tableProps
TypeSerializer<T> typeSerializer
Comparable<T> consumeStartOffset
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
CachedSerializedValue<T> sd
Properties properties
JobConfWrapper jobConfWrapper
List<E> partitionKeys
String[] fieldNames
DataType[] fieldTypes
String hiveVersion
HiveShim shim
RowType producedRowType
boolean useMapRedReader
JobConfWrapper jobConfWrapper
String hiveVersion
ObjectPath tablePath
DataType[] fieldTypes
String[] fieldNames
List<E> partitionKeys
int[] selectedFields
boolean useMapRedReader
ObjectPath tablePath
HiveShim hiveShim
JobConfWrapper confWrapper
List<E> partitionKeys
DataType[] fieldTypes
String[] fieldNames
Configuration configuration
String defaultPartitionName
ConsumeOrder consumeOrder
HiveTablePartition hiveTablePartition
HiveTableInputFormat inputFormat
HiveTablePartition hiveTablePartition
HiveTablePartition hiveTablePartition
HiveWriterFactory factory
HiveWriterFactory factory
Class<T> hiveOutputFormatClz
CachedSerializedValue<T> serDeInfo
String[] allColumns
DataType[] allTypes
String[] partitionColumns
Properties tableProperties
JobConfWrapper confWrapper
HiveShim hiveShim
boolean isCompressed
Boolean useManagedMemory
MemorySize fixedMemoryPerSlot
Double writeBufferRatio
Double highPriorityPoolRatio
StateBackend checkpointStreamBackend
File[] localRocksDbDirectories
PredefinedOptions predefinedOptions
RocksDBOptionsFactory rocksDbOptionsFactory
TernaryBoolean enableIncrementalCheckpointing
int numberOfTransferThreads
RocksDBMemoryConfiguration memoryConfiguration
RocksDBStateBackend.PriorityQueueStateType priorityQueueStateType
RocksDBNativeMetricOptions defaultMetricOptions
long writeBatchSize
RocksDBWriteBatchWrapper
, default value
2mb.Path file
long start
long length
URI uri
int partitionNumber
int totalNumberOfPartitions
int splitNumber
String[] hostnames
SerializableSupplier<T> serializerSupplier
byte[] buffer
int end
int position
int id
double x
double y
Collection<E> centroids
List<E> tasks
FileCopyTaskInputSplit curInputSplit
FileCopyTask task
int splitNumber
EnumTrianglesDataTypes.Edge outEdge
double dampening
double randomJump
double x
double y
double theta0
double theta1
Collection<E> parameters
LinearRegression.Params parameter
int count
EmptyFieldsCountAccumulator.VectorAccumulator emptyFieldCounter
int inputCase_
Object input_
byte memoizedIsInitialized
int windowType_
long lowerBoundary_
long upperBoundary_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int precision_
int scale_
byte memoizedIsInitialized
Object name_
Object description_
FlinkFnApi.Schema.FieldType type_
byte memoizedIsInitialized
int typeInfoCase_
Object typeInfo_
int typeName_
boolean nullable_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType keyType_
FlinkFnApi.Schema.FieldType valueType_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
Object name_
Object description_
FlinkFnApi.TypeInfo.FieldType type_
byte memoizedIsInitialized
int typeInfoCase_
Object typeInfo_
int typeName_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType elementType_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType keyType_
FlinkFnApi.Schema.FieldType valueType_
byte memoizedIsInitialized
List<E> udfs_
boolean metricEnabled_
com.google.protobuf.Internal.IntList grouping_
int groupingMemoizedSerializedSize
boolean generateUpdateBefore_
FlinkFnApi.Schema.FieldType keyType_
int indexOfCountStar_
boolean stateCleaningEnabled_
int stateCacheSize_
int mapStateReadCacheSize_
int mapStateWriteCacheSize_
boolean countStarInserted_
byte memoizedIsInitialized
int functionType_
FlinkFnApi.UserDefinedDataStreamFunction.RuntimeContext runtimeContext_
com.google.protobuf.ByteString payload_
boolean metricEnabled_
FlinkFnApi.TypeInfo keyTypeInfo_
byte memoizedIsInitialized
Object taskName_
Object taskNameWithSubtasks_
int numberOfParallelSubtasks_
int maxNumberOfParallelSubtasks_
int indexOfThisSubtask_
int attemptNumber_
List<E> jobParameters_
byte memoizedIsInitialized
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
DeserializationSchema<T> nestedSchema
GenericRecord
. *TypeInformation<T> typeInfo
AvroToRowDataConverters.AvroToRowDataConverter runtimeConverter
SerializationSchema<T> nestedSchema
GenericRecord
into bytes. *RowType rowType
RowDataToAvroConverters.RowDataToAvroConverter runtimeConverter
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
AvroBuilder<T> avroBuilder
DataFileWriter
.SchemaCoder.SchemaCoderProvider schemaCoderProvider
SchemaCoder.SchemaCoderProvider schemaCoderProvider
SchemaCoder schemaCoder
AvroRowDataDeserializationSchema avroDeserializer
TypeInformation<T> producedTypeInfo
RowData
.AvroRowDataSerializationSchema avroSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream ois) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
RowType formatRowType
DataType[] fieldTypes
String[] fieldNames
int[] selectFields
List<E> partitionKeys
String defaultPartValue
long limit
int[] csvSelectFieldToProjectFieldMapping
int[] csvSelectFieldToCsvFieldMapping
boolean ignoreParseErrors
TypeInformation<T> resultTypeInfo
CsvToRowDataConverters.CsvToRowDataConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader objectReader
CsvSchema
.boolean ignoreParseErrors
RowType rowType
RowDataToCsvConverters.RowDataToCsvConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvMapper csvMapper
JsonNode
into bytes.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectWriter objectWriter
CsvSchema
.TypeInformation<T> typeInfo
org.apache.flink.formats.csv.CsvRowDeserializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader objectReader
CsvSchema
.boolean ignoreParseErrors
RowTypeInfo typeInfo
org.apache.flink.formats.csv.CsvRowSerializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvMapper csvMapper
JsonNode
into bytes.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectWriter objectWriter
CsvSchema
.boolean ignoreParseErrors
TypeInformation<T>[] fieldTypes
String[] selectedFieldNames
boolean ignoreParseErrors
Path[] filePaths
TypeInformation<T>[] fieldTypes
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
boolean ignoreParseErrors
int[] selectedFields
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
boolean failOnMissingField
boolean ignoreParseErrors
TypeInformation<T> resultTypeInfo
RowData
.JsonToRowDataConverters.JsonToRowDataConverter runtimeConverter
JsonNode
s into objects of Flink SQL internal data
structures.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
TimestampFormat timestampFormat
RowType rowType
RowDataToJsonConverters.RowDataToJsonConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
TimestampFormat timestampFormat
JsonOptions.MapNullKeyMode mapNullKeyMode
String mapNullKeyLiteral
RowTypeInfo typeInfo
boolean failOnMissingField
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
org.apache.flink.formats.json.JsonRowDeserializationSchema.DeserializationRuntimeConverter runtimeConverter
boolean ignoreParseErrors
RowTypeInfo typeInfo
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
org.apache.flink.formats.json.JsonRowSerializationSchema.SerializationRuntimeConverter runtimeConverter
boolean failOnMissingField
boolean ignoreParseErrors
TimestampFormat timestampFormat
TimestampFormat timestampFormat
JsonOptions.MapNullKeyMode mapNullKeyMode
String mapNullKeyLiteral
JsonRowDataDeserializationSchema jsonDeserializer
TypeInformation<T> resultTypeInfo
RowData
. *String database
String table
boolean ignoreParseErrors
int fieldCount
JsonRowDataSerializationSchema jsonSerializer
JsonRowDataDeserializationSchema jsonDeserializer
boolean hasMetadata
org.apache.flink.formats.json.debezium.DebeziumJsonDeserializationSchema.MetadataConverter[] metadataConverters
TypeInformation<T> producedTypeInfo
TypeInformation
of the produced RowData
(physical + meta data).boolean schemaInclude
boolean ignoreParseErrors
JsonRowDataSerializationSchema jsonSerializer
JsonRowDataDeserializationSchema jsonDeserializer
TypeInformation<T> resultTypeInfo
RowData
. *boolean ignoreParseErrors
int fieldCount
JsonRowDataSerializationSchema jsonSerializer
TimestampFormat timestampFormat
RowType producedType
boolean skipWrongSchemaFileSplit
boolean skipCorruptedRecord
boolean skipThisSplit
TypeInformation<T>[] fieldTypes
String[] fieldNames
org.apache.parquet.filter2.predicate.FilterPredicate filterPredicate
Class<T> pojoTypeClass
TypeSerializer<T> typeSerializer
SerializableConfiguration hadoopConfig
String[] projectedFields
LogicalType[] projectedTypes
ColumnBatchFactory<SplitT extends FileSourceSplit> batchFactory
int batchSize
boolean isUtcTimestamp
boolean isCaseSensitive
ParquetBuilder<T> writerBuilder
RowType rowType
SerializableConfiguration configuration
boolean utcTimestamp
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration conf
org.apache.flink.formats.sequencefile.SerializableHadoopConfiguration serializableHadoopConfig
Class<T> keyClass
Class<T> valueClass
String compressionCodecName
org.apache.hadoop.io.SequenceFile.CompressionType compressionType
String id
long count
long checksum
MurmurHash hasher
Object vertexId0
long offset
double x
double y
double beta
long srcId
long offset
long length
long numberOfVertices
IterationRuntimeContext runtimeContext
long numberOfVertices
IterationRuntimeContext runtimeContext
TypeInformation<T> typeInformation
TypeInformation<T> typeInformation
Either.Right<L,R> nullValue
ByteValue bitmask
DoubleValue hubScore
DoubleValue authorityScore
DoubleValue pageRankScore
FloatValue adamicAdarScore
Vertex<K,V> outVertex
Tuple2<T0,T1> outMsg
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
org.apache.flink.graph.pregel.ComputeFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
boolean setNewVertexValueCalled
Object first
long numberOfVertices
boolean optDegrees
IterationRuntimeContext runtimeContext
Collector<T> out
Collector<T> outWithDegrees
Vertex<K,V> outVal
Vertex<K,V> outValWithDegrees
long inDegree
long outDegree
boolean setNewVertexValueCalled
long numberOfVertices
EdgeDirection direction
Tuple2<T0,T1> outValue
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
Object vertexId
org.apache.flink.graph.spargel.ScatterFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
long inDegree
long outDegree
boolean isBounded
byte[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.ByteValueArray.ReadIterator iterator
boolean ascendingComparison
ByteValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
char[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.CharValueArray.ReadIterator iterator
boolean ascendingComparison
CharValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
double[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.DoubleValueArray.ReadIterator iterator
boolean ascendingComparison
DoubleValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
float[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.FloatValueArray.ReadIterator iterator
boolean ascendingComparison
FloatValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
int[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.IntValueArray.ReadIterator iterator
boolean ascendingComparison
IntValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
long[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.LongValueArray.ReadIterator iterator
boolean ascendingComparison
LongValueArray reference
TypeComparator<T>[] comparators
int position
IntValue hashValue
org.apache.flink.graph.types.valuearray.NullValueArray.ReadIterator iterator
boolean ascendingComparison
NullValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
short[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.ShortValueArray.ReadIterator iterator
boolean ascendingComparison
ShortValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
int boundedLength
byte[] data
int length
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.StringValueArray.ReadIterator iterator
boolean ascendingComparison
StringValueArray reference
TypeComparator<T>[] comparators
TypeInformation<T> valueType
Class<T> type
Object value
int seed
int count
int hash
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
TypeSerializer<T> keySerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration configuration
org.apache.hive.hcatalog.mapreduce.HCatInputFormat hCatInputFormat
org.apache.hadoop.mapreduce.RecordReader<KEYIN,VALUEIN> recordReader
boolean fetched
boolean hasNext
String[] fieldNames
org.apache.hive.hcatalog.data.schema.HCatSchema outputSchema
TypeInformation<T> resultType
MesosWorkerStore.Worker worker
org.apache.mesos.Protos.TaskID taskID
scala.Option<A> slaveID
scala.Option<A> hostname
MesosWorkerStore.WorkerState state
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
scala.collection.Seq<A> tasks
scala.collection.Seq<A> newLeases
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
String hostname
private Object readResolve()
private Object readResolve()
scala.collection.Seq<A> tasks
boolean replace
private Object readResolve()
scala.collection.immutable.Map<A,B> remaining
int retries
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.SlaveID slaveID
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.SlaveID slaveID
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.mesos.scheduler.TaskMonitor.TaskGoalState goal
private Object readResolve()
private Object readResolve()
org.apache.flink.mesos.scheduler.TaskMonitor.TaskGoalState state
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.TaskStatus status
private Object readResolve()
String hostname
Collection<E> offerIds
Collection<E> operations
org.apache.mesos.Protos.Filters filters
String message
org.apache.mesos.Protos.ExecutorID executorId
org.apache.mesos.Protos.SlaveID slaveId
int status
org.apache.mesos.Protos.ExecutorID executorId
org.apache.mesos.Protos.SlaveID slaveId
byte[] data
org.apache.mesos.Protos.OfferID offerId
org.apache.mesos.Protos.FrameworkID frameworkId
org.apache.mesos.Protos.MasterInfo masterInfo
org.apache.mesos.Protos.MasterInfo masterInfo
org.apache.mesos.Protos.SlaveID slaveId
org.apache.mesos.Protos.TaskStatus status
int m
int n
double[] data
double[] data
int n
int[] indices
double[] values
Mapper mapper
ModelMapper mapper
ModelSource modelSource
String modelVariableName
String[] inputColNames
TypeInformation<T>[] inputColTypes
String[] outputColNames
TypeInformation<T>[] outputColTypes
int[] reservedCols
int[] reservedColsPosInResult
int[] outputColsPosInResult
Table modelData
int fieldNumber
OrcShim<BATCH> shim
SerializableHadoopConfigWrapper hadoopConfigWrapper
org.apache.orc.TypeDescription schema
int[] selectedFields
List<E> conjunctPredicates
int batchSize
ColumnBatchFactory<BatchT,SplitT extends FileSourceSplit> batchFactory
RowType projectedOutputType
Serializable lowerBound
Serializable upperBound
Serializable[] literals
OrcFilters.Predicate pred
OrcFilters.Predicate[] preds
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
int batchSize
Configuration conf
org.apache.orc.TypeDescription schema
int[] selectedFields
ArrayList<E> conjunctPredicates
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration conf
String schema
LogicalType[] fieldTypes
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
LogicalType[] fieldTypes
org.apache.orc.TypeDescription schema
Vectorizer<T> vectorizer
Properties writerProperties
Map<K,V> confMap
org.apache.orc.OrcFile.WriterOptions writerOptions
int maxBundleSize
long maxBundleTimeMills
int maxArrowBatchSize
Map<K,V> pythonFilesInfo
String pythonRequirementsFileInfo
String pythonRequirementsCacheDirInfo
Map<K,V> pythonArchivesInfo
String pythonExec
boolean metricEnabled
boolean isUsingManagedMemory
Configuration mergedConfig
OperatorID generatedOperatorID
OperatorID userDefinedOperatorID
JobID jobID
ExecutionAttemptID executionAttemptID
SerializedValue<T> userAccumulators
byte[] key
org.apache.flink.runtime.blob.BlobKey.BlobType type
AbstractID random
long checkpointId
long triggerTimestamp
Map<K,V> taskStats
TaskStateStats
accessible by their ID.int numberOfSubtasks
CheckpointProperties props
CheckpointFailureReason checkpointFailureReason
long checkpointId
long timestamp
long receiveTimestamp
long bytesProcessedDuringAlignment
long bytesPersistedDuringAlignment
long alignmentDurationNanos
long syncDurationMillis
long asyncDurationMillis
long checkpointStartDelayNanos
CheckpointType checkpointType
CheckpointStorageLocationReference targetLocation
CheckpointOptions.AlignmentType alignmentType
long alignmentTimeout
CheckpointType checkpointType
boolean forced
boolean discardSubsumed
boolean discardFinished
boolean discardCancelled
boolean discardFailed
boolean discardSuspended
AtomicInteger numberOfCheckpointsToClean
long numRestoredCheckpoints
long numTotalCheckpoints
int numInProgressCheckpoints
long numCompletedCheckpoints
long numFailedCheckpoints
List<E> checkpointsHistory
CheckpointStatsHistory.createSnapshot()
.Map<K,V> checkpointsById
CheckpointStatsHistory.createSnapshot()
.int maxSize
boolean readOnly
CompletedCheckpointStats latestCompletedCheckpoint
FailedCheckpointStats latestFailedCheckpoint
CompletedCheckpointStats latestSavepoint
CheckpointStatsCounts counts
CompletedCheckpointStatsSummary summary
CheckpointStatsHistory history
RestoredCheckpointStats latestRestoredCheckpoint
JobID job
long checkpointID
long timestamp
long duration
Map<K,V> operatorStates
CheckpointProperties props
Collection<E> masterHookStates
CompletedCheckpointStorageLocation storageLocation
StreamStateHandle metadataHandle
String externalPointer
long stateSize
long processedData
long persistedData
SubtaskStateStats latestAcknowledgedSubtask
String externalPointer
boolean discarded
MinMaxAvgStats stateSize
MinMaxAvgStats duration
MinMaxAvgStats processedData
MinMaxAvgStats persistedData
boolean isDiscarded
boolean registered
boolean discarded
int numAcknowledgedSubtasks
long stateSize
long processedData
long persistedData
long failureTimestamp
SubtaskStateStats latestAcknowledgedSubtask
String failureMsg
long restoreCheckpointId
TaskStateSnapshot taskStateSnapshot
String name
byte[] bytes
int version
long min
long max
long sum
long count
OperatorID operatorID
Map<K,V> operatorSubtaskStates
ByteStreamStateHandle coordinatorState
int parallelism
int maxParallelism
StateObjectCollection<T extends StateObject> managedOperatorState
OperatorStateBackend
.StateObjectCollection<T extends StateObject> rawOperatorState
OperatorStateCheckpointOutputStream
.StateObjectCollection<T extends StateObject> managedKeyedState
KeyedStateBackend
.StateObjectCollection<T extends StateObject> rawKeyedState
KeyedStateCheckpointOutputStream
.StateObjectCollection<T extends StateObject> inputChannelState
StateObjectCollection<T extends StateObject> resultSubpartitionState
InflightDataRescalingDescriptor inputRescalingDescriptor
StateAssignmentOperation
and will not be persisted in the checkpoint itself as it can only
be calculated if the the post-recovery scale factor is known.InflightDataRescalingDescriptor outputRescalingDescriptor
StateAssignmentOperation
and will not
be persisted in the checkpoint itself as it can only be calculated if the the post-recovery
scale factor is known.long stateSize
int currentNumAcknowledgedSubtasks
long currentStateSize
long currentProcessedData
long currentPersistedData
SubtaskStateStats latestAcknowledgedSubtask
long checkpointId
CheckpointProperties props
long restoreTimestamp
String externalPath
Collection<E> stateObjects
ChainedStateHandle<T extends StateObject> managedOperatorState
OperatorStateBackend
.ChainedStateHandle<T extends StateObject> rawOperatorState
OperatorStateCheckpointOutputStream
.KeyedStateHandle managedKeyedState
KeyedStateBackend
.KeyedStateHandle rawKeyedState
KeyedStateCheckpointOutputStream
.long stateSize
int subtaskIndex
long ackTimestamp
long stateSize
long syncCheckpointDuration
long asyncCheckpointDuration
long processedData
long persistedData
long alignmentDuration
long checkpointStartDelay
JobVertexID jobVertexID
Map<K,V> subtaskStates
int parallelism
int maxParallelism
int chainLength
JobVertexID jobVertexId
SubtaskStateStats[] subtaskStats
TaskStateStats.TaskStateStatsSummary summaryStats
int numAcknowledgedSubtasks
SubtaskStateStats latestAckedSubtaskStats
MinMaxAvgStats stateSize
MinMaxAvgStats ackTimestamp
MinMaxAvgStats syncCheckpointDuration
MinMaxAvgStats asyncCheckpointDuration
MinMaxAvgStats processedData
MinMaxAvgStats persistedData
MinMaxAvgStats alignmentDuration
MinMaxAvgStats checkpointStartDelay
int gateIdx
int inputChannelIdx
int partitionIdx
int subPartitionIdx
JobID jobID
HashMap<K,V> taskManagerEnv
TaskExecutorProcessSpec taskExecutorProcessSpec
Configuration systemProperties
List<E> artifacts
Map<K,V> environmentVariables
Configuration flinkConfiguration
CPUResource cpuCores
private Object readResolve()
Resource cpuCores
MemorySize taskHeapMemory
MemorySize taskOffHeapMemory
MemorySize managedMemory
MemorySize networkMemory
Map<K,V> extendedResources
ResourceSpec
.ResourceID resourceId
int slotNumber
IntermediateDataSetID consumedResultId
ResultPartitionType consumedPartitionType
int consumedSubpartitionIndex
DistributionPattern
and the subtask indices of the producing and consuming task.ShuffleDescriptor[] inputChannels
PartitionDescriptor partitionDescriptor
ShuffleDescriptor shuffleDescriptor
int maxParallelism
boolean sendScheduleOrUpdateConsumersMessage
TaskDeploymentDescriptor.MaybeOffloaded<T> serializedJobInformation
TaskDeploymentDescriptor.MaybeOffloaded<T> serializedTaskInformation
JobID jobId
NOTE: this is redundant to the information stored in TaskDeploymentDescriptor.serializedJobInformation
but
needed in order to restore offloaded data.
ExecutionAttemptID executionId
AllocationID allocationId
int subtaskIndex
int attemptNumber
List<E> producedPartitions
List<E> inputGates
int targetSlotNumber
JobManagerTaskRestore taskRestore
SerializedValue<T> serializedValue
PermanentBlobKey serializedValueKey
String blobServerHostname
int blobServerPort
ExecutionAttemptID attemptId
long[] stateTimestamps
int attemptNumber
ExecutionState state
String failureCause
TaskManagerLocation assignedResourceLocation
AllocationID assignedAllocationID
StringifiedAccumulatorResult[] userAccumulators
int parallelSubtaskIndex
IOMetrics ioMetrics
JobID jobID
String jobName
Map<K,V> tasks
List<E> verticesInCreationOrder
long[] stateTimestamps
System.currentTimeMillis()
when the
execution graph transitioned into a certain state. The index into this array is the ordinal
of the enum value, i.e. the timestamp when the graph went into state "RUNNING" is at stateTimestamps[RUNNING.ordinal()]
.JobStatus state
ErrorInfo failureCause
String jsonPlan
StringifiedAccumulatorResult[] archivedUserAccumulators
ArchivedExecutionConfig archivedExecutionConfig
boolean isStoppable
Map<K,V> serializedUserAccumulators
CheckpointCoordinatorConfiguration jobCheckpointingConfiguration
CheckpointStatsSnapshot checkpointStatsSnapshot
String stateBackendName
ArchivedExecutionVertex[] taskVertices
JobVertexID id
String name
int parallelism
int maxParallelism
ResourceProfile resourceProfile
StringifiedAccumulatorResult[] archivedUserAccumulators
int subTaskIndex
EvictingBoundedList<T> priorExecutions
String taskNameWithSubtask
ArchivedExecution currentExecution
SerializedThrowable exception
long timestamp
AbstractID executionAttemptId
long expectedModVersion
long actualModVersion
long numRecordsIn
long numRecordsOut
long numBytesIn
long numBytesOut
JobID jobId
String jobName
SerializedValue<T> serializedExecutionConfig
Configuration jobConfiguration
Collection<E> requiredJarFileBlobKeys
Collection<E> requiredClasspathURLs
IntermediateDataSetID intermediateDataSetID
ShuffleDescriptor shuffleDescriptor
JobVertexID jobVertexId
String taskName
int numberOfSubtasks
int maxNumberOfSubtasks
String invokableClassName
Configuration taskConfiguration
int numberOfFailures
int maxNumberRestartAttempts
long delayBetweenRestartAttempts
int numberOfCPUCores
long sizeOfPhysicalMemory
long sizeOfJvmHeap
long sizeOfManagedMemory
boolean closed
InetSocketAddress address
int connectionIndex
SocketAddress address
ResultPartitionID partitionId
IntermediateResultPartitionID partitionId
ExecutionAttemptID producerId
IntermediateDataSetID consumedResultID
ExecutionAttemptID consumerID
IntermediateDataSetID id
JobVertex producer
List<E> consumers
ResultPartitionType resultType
IntermediateDataSetID intermediateDataSetID
int partitionNum
JobVertex target
DistributionPattern distributionPattern
SubtaskStateMapper downstreamSubtaskStateMapper
SubtaskStateMapper upstreamSubtaskStateMapper
IntermediateDataSet source
IntermediateDataSetID sourceId
String shipStrategyName
String preProcessingOperationName
String operatorLevelCachingDescription
Map<K,V> taskVertices
Configuration jobConfiguration
JobID jobID
String jobName
ScheduleMode scheduleMode
boolean approximateLocalRecovery
SerializedValue<T> serializedExecutionConfig
JobCheckpointingSettings snapshotSettings
SavepointRestoreSettings savepointRestoreSettings
List<E> userJars
Map<K,V> userArtifacts
List<E> userJarBlobKeys
List<E> classpaths
JobVertexID id
List<E> operatorIDs
The ID pairs are stored depth-first post-order; for the forking chain below the ID's would
be stored as [D, E, B, C, A]. A - B - D \ \ C E This is the same order that operators are
stored in the StreamTask
.
ArrayList<E> results
ArrayList<E> inputs
ArrayList<E> operatorCoordinators
int parallelism
int maxParallelism
ResourceSpec minResources
ResourceSpec preferredResources
Configuration configuration
String invokableClassName
boolean isStoppable
InputSplitSource<T extends InputSplit> inputSplitSource
String name
SlotSharingGroup slotSharingGroup
CoLocationGroup coLocationGroup
String operatorName
String operatorDescription
String operatorPrettyName
String resultOptimizerProperties
InputDependencyConstraint inputDependencyConstraint
String restorePath
boolean allowNonRestoredState
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
int tolerableCheckpointFailureNumber
CheckpointRetentionPolicy checkpointRetentionPolicy
boolean isExactlyOnce
false
, at least once mode has been configured. This is not a necessary attribute, because
the checkpointing mode is only relevant for the stream tasks, but we expose it here to
forward it to the web runtime UI.boolean isPreferCheckpointForRecovery
boolean isUnalignedCheckpointsEnabled
long alignmentTimeout
List<E> verticesToTrigger
List<E> verticesToAcknowledge
List<E> verticesToConfirm
CheckpointCoordinatorConfiguration checkpointCoordinatorConfiguration
SerializedValue<T> defaultStateBackend
SerializedValue<T> masterHooks
AbstractID id
List<E> vertices
Set<E> ids
SlotSharingGroupId slotSharingGroupId
ResourceSpec resourceSpec
int slotIndex
AllocationID allocationId
JobID jobId
Collection<E> allocatedSlotInfos
String reason
ResourceID resourceID
ResourceManagerId resourceManagerId
ResourceID resourceManagerResourceId
JobID jobId
ApplicationStatus applicationStatus
Map<K,V> accumulatorResults
long netRuntime
SerializedThrowable serializedThrowable
null
if the job finished successfully.byte[] inputSplitData
protected Object readResolve() throws ObjectStreamException
ObjectStreamException
int requestId
ExecutionAttemptID executionAttemptID
double backPressureRatio
JobID job
ExecutionAttemptID taskExecutionId
long checkpointId
TaskStateSnapshot subtaskState
CheckpointMetrics checkpointMetrics
SerializedCheckpointException serializedCheckpointException
CheckpointFailureReason checkpointFailureReason
SerializedThrowable serializedThrowable
int numTaskManagersConnected
int numSlotsTotal
int numSlotsAvailable
Collection<E> jobsWithStatus
int numJobsRunningOrPending
int numJobsFinished
int numJobsCancelled
int numJobsFailed
Collection<E> jobs
boolean includeRunning
boolean includeFinished
private Object readResolve()
private Object readResolve()
private Object readResolve()
byte[] serializedCounters
byte[] serializedGauges
byte[] serializedMeters
byte[] serializedHistograms
int numCounters
int numGauges
int numMeters
int numHistograms
int cnt
int cnt
Record res
int cnt
IntValue theInteger
int cnt
int cnt
int cnt
StringBuilder bld
int cnt
int cnt
boolean closed
int cnt
OperatorID operatorID
OperatorID operatorId
Object payload
Object payload
OperatorID operatorId
org.apache.flink.runtime.util.SerializableFunction<T,R> factory
int value
Integer count
boolean opened
boolean closed
boolean opened
boolean closed
int val
int size
int val
boolean isLong
int val
TypeComparatorFactory<T> typeComparator
int parallelism
TypeComparatorFactory<T> comparatorFactory
Configuration config
int value
String value
int stateReference
Long value
int numberOfDiscardCalls
JobID jobId
JobVertexID jobVertexId
int numKeyGroups
String registrationName
KvStateID[] kvStateIds
InetSocketAddress[] kvStateAddresses
int numRegisteredKeyGroups
SerializedThrowable reason
int numberTaskManagers
int numberRegisteredSlots
int numberFreeSlots
ResourceProfile totalResource
ResourceProfile freeResource
JobID jobId
AllocationID allocationId
ResourceProfile resourceProfile
String targetAddress
String taskExecutorAddress
ResourceID resourceId
int dataPort
int jmxPort
HardwareDescription hardwareDescription
TaskExecutorMemoryConfiguration memoryConfiguration
ResourceProfile defaultSlotResourceProfile
ResourceProfile totalResourceProfile
CPUResource cpuCores
MemorySize taskHeapSize
MemorySize taskOffHeapSize
MemorySize networkMemSize
MemorySize managedMemSize
int responseCode
RestHandlerException.LoggingBehavior loggingBehavior
int requestId
long endTimestamp
double[] subTaskBackPressureRatios
double maxSubTaskBackPressureRatio
OperatorBackPressureStats operatorBackPressureStats
boolean numBytesInComplete
boolean numBytesOutComplete
boolean numRecordsInComplete
boolean numRecordsOutComplete
String name
long size
Collection<E> logInfos
JobIDDeserializer jobIdDeserializer
SerializedThrowableDeserializer serializedThrowableDeserializer
SerializedValueDeserializer serializedValueDeserializer
JobIDSerializer jobIdSerializer
SerializedValueSerializer serializedValueSerializer
SerializedThrowableSerializer serializedThrowableSerializer
TaskManagerMetricsInfo taskManagerMetrics
ResourceID resourceId
String address
int dataPort
int jmxPort
long lastHeartbeat
int numberSlots
int numberAvailableSlots
ResourceProfileInfo totalResource
ResourceProfileInfo freeResource
HardwareDescription hardwareDescription
TaskExecutorMemoryConfiguration memoryConfiguration
Collection<E> threadInfos
int responseCode
Serializable fencingToken
Serializable payload
private void readObject(ObjectInputStream ois) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
SerializedValue<T> serializedMethodInvocation
JobVertexID jobVertexId
int subtaskIndex
ResourceID producerLocation
NettyShuffleDescriptor.PartitionConnectionInfo partitionConnectionInfo
ResultPartitionID resultPartitionID
ConnectionID connectionID
IntermediateDataSetID resultId
int totalNumberOfPartitions
IntermediateResultPartitionID partitionId
ResultPartitionType partitionType
int numberOfSubpartitions
int connectionIndex
ResultPartitionID resultPartitionID
ResourceProfile resourceProfile
int numberOfRequiredSlots
JobID jobId
String targetAddress
Collection<E> resourceRequirements
int subtaskId
String location
String hostName
SourceEvent sourceEvent
Object info
StreamStateHandle delegate
List<E> offsets
long size
int subtaskIndex
TypeSerializer<T> elementSerializer
protected final Object readResolve() throws ObjectStreamException
ObjectStreamException
byte[] encodedReference
DirectoryStateHandle directoryStateHandle
KeyGroupRange keyGroupRange
String directoryString
long checkpointId
UUID backendIdentifier
StreamStateHandle metaDataState
Set<E> sharedStateHandleIDs
UUID backendIdentifier
SharedStateRegistry
.KeyGroupRange keyGroupRange
long checkpointId
Map<K,V> sharedState
Map<K,V> privateState
StreamStateHandle metaStateHandle
int startKeyGroup
int endKeyGroup
KeyGroupRange keyGroupRange
long[] offsets
KeyGroupRangeOffsets groupRangeOffsets
StreamStateHandle stateHandle
File[] allocationBaseDirs
JobID jobID
JobVertexID jobVertexID
int subtaskIndex
long[] offsets
OperatorStateHandle.Mode distributionMode
Map<K,V> stateNameToPartitionOffsets
StreamStateHandle delegateStateHandle
StreamStateHandle wrappedStreamStateHandle
StateObject jobManagerOwnedSnapshot
StateObject taskLocalSnapshot
boolean disposed
Path filePath
long stateSize
Path exclusiveCheckpointDir
FileStateHandle metadataFileHandle
String externalPointer
int fileStateThreshold
TernaryBoolean asynchronousSnapshots
int writeBufferSize
String relativePath
boolean disabled
byte[] data
String handleName
ByteStreamStateHandle
with the exact same name must also have the exact same content
in data.int maxStateSize
TernaryBoolean asynchronousSnapshots
Object userValue
long lastAccessTimestamp
Collection<E> accumulatorSnapshots
Collection<E> slotsStatus
SlotID slotID
ResourceProfile resourceProfile
AllocationID allocationID
JobID jobID
SlotReport slotReport
ClusterPartitionReport clusterPartitionReport
Long frameworkHeap
Long taskHeap
Long frameworkOffHeap
Long taskOffHeap
Long networkMemory
Long managedMemoryTotal
Long jvmMetaspace
Long jvmOverhead
Long totalFlinkMemory
Long totalProcessMemory
String reason
InstanceID registrationId
ResourceID resourceManagerResourceId
ClusterInformation clusterInformation
AccumulatorReport accumulatorReport
ExecutionDeploymentReport executionDeploymentReport
AllocationID allocationId
JobID jobId
Collection<E> entries
IntermediateDataSetID dataSetId
Set<E> hostedPartitions
int numTotalPartitions
AllocationID allocationId
int slotIndex
ResourceProfile resourceProfile
JobID jobID
ExecutionAttemptID executionId
ExecutionState executionState
SerializedThrowable throwable
AccumulatorSnapshot accumulators
IOMetrics ioMetrics
ResourceID resourceID
InetAddress inetAddress
TaskManagerLocation.HostNameSupplier hostNameSupplier
int dataPort
String stringRepresentation
InetAddress inetAddress
String hostName
String fqdnHostName
InetAddress inetAddress
ResourceID resourceID
String externalAddress
int dataPort
FlinkMemory flinkMemory
JvmMetaspaceAndOverhead jvmMetaspaceAndOverhead
MemorySize metaspace
MemorySize overhead
MemorySize jvmHeap
MemorySize offHeapMemory
MemorySize frameworkHeap
MemorySize frameworkOffHeap
MemorySize taskHeap
MemorySize taskOffHeap
MemorySize network
MemorySize managed
org.apache.calcite.sql.parser.SqlParserPos errorPosition
String message
boolean specialConstructor
Token currentToken
int[][] expectedTokenSequences
String[] tokenImage
String eol
int errorCode
boolean specialConstructor
Token currentToken
int[][] expectedTokenSequences
String[] tokenImage
String eol
int errorCode
int numKeys
int keyIndex
long durationMs
long stepMs
long offsetMs
long ms
MapStateDescriptor<UK,UV> descriptor
OperatorState operatorState
StateBackend stateBackend
Configuration configuration
StateReaderOperator<F extends Function,KEY,N,OUT> operator
ListStateDescriptor<T> descriptor
ListStateDescriptor<T> descriptor
Function function
TypeInformation<T> keyType
TypeSerializer<T> namespaceSerializer
WindowContents<S extends State,IN> contents
StateDescriptor<S extends State,T> descriptor
AggregateFunction<IN,ACC,OUT> aggFunction
ReduceFunction<T> reduceFunction
StreamConfig streamConfig
int maxParallelism
Timestamper<T> timestamper
String path
Collection<E> masterStates
OperatorID operatorID
int maxParallelism
Path savepointPath
TimestampAssigner<T> assigner
long timestamp
Path savepointPath
long timestamp
Path savepointPath
long timestamp
Path savepointPath
long timestamp
Path savepointPath
Output<T> output
AbstractStreamOperator<OUT> operator
KeySelector<IN,KEY> keySelector
int maxParallelism
TypeSerializer<T> oneSerializer
TypeSerializer<T> twoSerializer
CheckpointingMode checkpointingMode
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
boolean forceCheckpointing
boolean forceUnalignedCheckpoints
boolean unalignedCheckpointsEnabled
long alignmentTimeout
boolean approximateLocalRecovery
CheckpointConfig.ExternalizedCheckpointCleanup externalizedCheckpointCleanup
boolean failOnCheckpointingErrors
CheckpointConfig.tolerableCheckpointFailureNumber
.CheckpointConfig.tolerableCheckpointFailureNumber
would always overrule this deprecated field if
they have conflicts.
boolean preferCheckpointForRecovery
int tolerableCheckpointFailureNumber
CheckpointConfig.setTolerableCheckpointFailureNumber(int)
.long maxTimestamp
Comparator comparator
boolean byAggregate
boolean first
FieldAccessor<T,F> fieldAccessor
FieldAccessor<T,F> fieldAccessor
SumFunction adder
TypeSerializer<T> serializer
boolean isTuple
byte[] serializedPythonFunction
PythonEnv pythonEnv
PythonFunction pythonFunction
int functionType
boolean isKeyOfPickledByteArray
OutputFormat<IT> format
boolean cleanupCalled
PrintSinkOutputWriter<IN> writer
SerializableObject lock
SerializationSchema<T> schema
String hostName
int port
int maxNumRetries
boolean autoFlush
int retries
boolean isRunning
LinkedHashMap<K,V> pendingCommitTransactions
java.time.Clock clock
ListStateDescriptor<T> stateDescriptor
TwoPhaseCommitSinkFunction.TransactionHolder<TXN> currentTransactionHolder
long transactionTimeout
boolean ignoreFailuresAfterTransactionTimeout
TwoPhaseCommitSinkFunction.recoverAndCommit(Object)
will be caught instead of
propagated.double transactionTimeoutWarningRatio
TypeSerializer<T> transactionSerializer
TypeSerializer<T> contextSerializer
String path
ArrayList<E> tupleList
WriteFormat<IN> format
long millis
long lastTime
Path basePath
HadoopPathBasedBulkWriter.Factory<T> writerFactory
HadoopFileCommitterFactory fileCommitterFactory
SerializableConfiguration serializableConfiguration
BucketAssigner<IN,BucketID> bucketAssigner
CheckpointRollingPolicy<IN,BucketID> rollingPolicy
BucketFactory<IN,BucketID> bucketFactory
OutputFileConfig outputFileConfig
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
long bucketCheckInterval
StreamingFileSink.BucketsBuilder<IN,BucketID,T extends StreamingFileSink.BucketsBuilder<IN,BucketID,T>> bucketsBuilder
long bucketCheckInterval
Path basePath
BulkWriter.Factory<T> writerFactory
BucketAssigner<IN,BucketID> bucketAssigner
CheckpointRollingPolicy<IN,BucketID> rollingPolicy
BucketFactory<IN,BucketID> bucketFactory
OutputFileConfig outputFileConfig
long bucketCheckInterval
Path basePath
Encoder<IN> encoder
BucketAssigner<IN,BucketID> bucketAssigner
RollingPolicy<IN,BucketID> rollingPolicy
BucketFactory<IN,BucketID> bucketFactory
OutputFileConfig outputFileConfig
String formatString
java.time.ZoneId zoneId
long partSize
long rolloverInterval
long inactivityInterval
String path
int readerParallelism
FileInputFormat<OT> format
FileInputFormat
to be read.long interval
FileProcessingMode watchType
FileProcessingMode
.long globalModificationTime
boolean isRunning
TypeSerializer<T> serializer
InputFormat<OT,T extends InputSplit> inputFormat
TypeInformation<T> type
ExecutionConfig executionConfig
TypeSerializer<T> serializer
byte[] elementsSerialized
int numElements
int numElementsEmitted
int numElementsToSkip
boolean isRunning
SplittableIterator<T> fullIterator
boolean isRunning
TypeInformation<T> typeInfo
InputFormat<OT,T extends InputSplit> format
boolean isRunning
TypeSerializer<T> idSerializer
long start
long end
boolean isRunning
long modificationTime
Serializable splitState
DataGenerator<T> generator
long rowsPerSecond
Long numberOfRows
long start
long end
long currentTimestamp
AscendingTimestampExtractor.MonotonyViolationHandler violationHandler
long currentMaxTimestamp
long lastEmittedWatermark
long maxOutOfOrderness
AggregateFunction<IN,ACC,OUT> aggFunction
AggregateFunction<IN,ACC,OUT> aggFunction
Window window
ProcessWindowFunction.Context context
ReduceFunction<T> reduceFunction
ReduceFunction<T> reduceFunction
ProcessAllWindowFunction<IN,OUT,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
ProcessWindowFunction<IN,OUT,KEY,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
int[] order
int fieldId
int fieldId
int[] indexes
Configuration config
TypeSerializer<T> typeSerializer
StreamConfig.InputRequirement inputRequirement
int inputGateIndex
StreamEdge inputEdge
String edgeId
int sourceId
int targetId
int typeNumber
OutputTag<T> outputTag
StreamEdge
.StreamPartitioner<T> outputPartitioner
StreamPartitioner
on this StreamEdge
.String sourceOperatorName
String targetOperatorName
ShuffleMode shuffleMode
long bufferTimeout
boolean supportsUnalignedCheckpoints
ChainingStrategy chainingStrategy
long combinedWatermark
long input1Watermark
long input2Watermark
ChainingStrategy chainingStrategy
StreamConfig config
Output<T> output
StreamingRuntimeContext runtimeContext
ExecutionConfig executionConfig
ClassLoader userCodeClassLoader
CloseableRegistry cancelables
long[] inputWatermarks
OperatorMetricGroup metrics
LatencyStats latencyStats
ProcessingTimeService processingTimeService
StreamOperatorStateHandler stateHandler
InternalTimeServiceManager<K> timeServiceManager
long combinedWatermark
Function userFunction
TypeSerializer<T> serializer
InternalTimerService<N> timerService
long inputMask
TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
long currentWatermark
InternalTimerService
.StreamSource<OUT,SRC extends SourceFunction<OUT>> operator
StreamOperator<OUT> operator
StreamSink<IN> operator
AbstractUdfStreamOperator<OUT,F extends Function> operator
FunctionWithException<T,R,E extends Throwable> readerFactory
SimpleVersionedSerializer<E> splitSerializer
OperatorEventGateway operatorEventGateway
WatermarkStrategy<T> watermarkStrategy
Configuration configuration
String localHostname
boolean emitProgressiveWatermarks
SourceReader<T,SplitT extends SourceSplit> sourceReader
ReaderOutput<T> currentMainOutput
PushingAsyncDataInput.DataOutput<T> lastInvokedOutput
ListState<T> readerState
TimestampsAndWatermarks<T> eventTimeLogic
boolean closed
Source<T,SplitT extends SourceSplit,EnumChkT> source
Source
to create the SourceOperator
.WatermarkStrategy<T> watermarkStrategy
boolean emitProgressiveWatermarks
int numCoordinatorWorkerThread
TypeSerializer<T> serializer
TypeSerializer<T> outSerializer
int[] fields
int numFields
long currentWatermark
InternalTimerService
.TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
int length
boolean immutableType
int capacity
AsyncDataStream.OutputMode outputMode
long timeout
AsyncFunction<IN,OUT> asyncFunction
long timeout
int capacity
AsyncDataStream.OutputMode outputMode
MailboxExecutor mailboxExecutor
long currentWatermark
InternalTimerService
.List<E> broadcastStateDescriptors
long currentWatermark
InternalTimerService
.long lowerBound
long upperBound
TypeSerializer<T> leftTypeSerializer
TypeSerializer<T> rightTypeSerializer
TypeSerializer<T> elementSerializer
String version
long offset
InetSocketAddress address
TypeSerializer<T> serializer
long maxBytesPerBatch
long bufferSizeLimitBytes
String accumulatorName
CollectSinkFunction<IN> sinkFunction
OperatorID operatorId
int socketTimeout
CollectSinkOperator<IN> operator
int socketTimeoutMillis
PythonConfig config
Map<K,V> jobOptions
TypeInformation<T> runnerInputTypeInfo
TypeInformation<T> runnerOutputTypeInfo
DataStreamPythonFunctionInfo pythonFunctionInfo
Map<K,V> jobOptions
TypeInformation<T> inputTypeInfo
TypeInformation<T> outputTypeInfo
DataStreamPythonFunctionInfo pythonFunctionInfo
int numPartitions
WatermarkStrategy<T> watermarkStrategy
TypeInformation<T> inputTypeInfo
boolean emitProgressiveWatermarks
Map<K,V> jobOptions
TypeInformation<T> inputTypeInfo1
TypeInformation<T> inputTypeInfo2
TypeInformation<T> outputTypeInfo
DataStreamPythonFunctionInfo pythonFunctionInfo
boolean isKeyedStream
byte[] data
int offset
int limit
long size
SessionWindowTimeGapExtractor<T> sessionWindowTimeGapExtractor
SessionWindowTimeGapExtractor<T> sessionWindowTimeGapExtractor
long sessionTimeout
long sessionTimeout
long size
long slide
long offset
long size
long offset
long slide
long size
long globalOffset
Long staggerOffset
WindowStagger windowStagger
long size
long globalOffset
Long staggerOffset
WindowStagger windowStagger
long maxCount
boolean doEvictAfter
DeltaFunction<DATA> deltaFunction
double threshold
boolean doEvictAfter
long windowSize
boolean doEvictAfter
long interval
ReducingStateDescriptor<T> stateDesc
long interval
ReducingStateDescriptor<T> stateDesc
long maxCount
ReducingStateDescriptor<T> stateDesc
DeltaFunction<DATA> deltaFunction
double threshold
ValueStateDescriptor<T> stateDesc
String insertQuery
boolean ignoreNullFields
ClusterBuilder builder
String keySpace
String table
Map<K,V> lastCommittedCheckpoints
CassandraCommitter.isCheckpointCommitted(int, long)
.Class<T> clazz
MapperOptions options
String keyspace
int rowArity
String insertQuery
ClusterBuilder builder
int arity
org.slf4j.Logger log
AtomicReference<V> throwable
com.google.common.util.concurrent.FutureCallback<V> callback
Semaphore semaphore
ClusterBuilder builder
CassandraSinkBaseConfig config
CassandraFailureHandler failureHandler
int maxConcurrentRequests
java.time.Duration maxConcurrentRequestsTimeout
boolean ignoreNullFields
String insertQuery
ClusterBuilder builder
Integer bulkProcessorFlushMaxActions
Integer bulkProcessorFlushMaxSizeMb
Long bulkProcessorFlushIntervalMillis
ElasticsearchSinkBase.BulkFlushBackoffPolicy bulkProcessorFlushBackoffPolicy
Map<K,V> userConfig
For TransportClient
based implementations, this
config map would also contain Elasticsearch-shipped configuration, and therefore this config
map would also be forwarded when creating the Elasticsearch client.
ElasticsearchSinkFunction<T> elasticsearchSinkFunction
ActionRequests
from
each incoming element.ActionRequestFailureHandler failureHandler
ActionRequests
.boolean flushOnCheckpoint
ElasticsearchApiCallBridge<C extends AutoCloseable> callBridge
AtomicLong numPendingRequests
ElasticsearchSinkBase.flushOnCheckpoint
is true
.
This is incremented whenever the user adds (or re-adds through the ActionRequestFailureHandler
) requests to the RequestIndexer
. It is decremented for
each completed request of a bulk request, in BulkProcessor.Listener#afterBulk(long,
BulkRequest, BulkResponse)
and BulkProcessor.Listener#afterBulk(long, BulkRequest,
Throwable)
.
AtomicReference<V> failureThrowable
BulkProcessor.Listener
if a Throwable
was thrown
in callbacks and the user considered it should fail the sink via the ActionRequestFailureHandler.onFailure(ActionRequest, Throwable, int, RequestIndexer)
method.
Errors will be checked and rethrown before processing each input element, and when the sink is closed.
ElasticsearchSinkBase.FlushBackoffType backoffType
int maxRetryCount
long delayMillis
IndexGenerator indexGenerator
String docType
String keyDelimiter
String keyNullLiteral
SerializationSchema<T> serializationSchema
org.elasticsearch.common.xcontent.XContentType contentType
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory
int[] keyFieldIndices
String dateTimeFormat
String index
String index
List<E> transportAddresses
We are using InetSocketAddress
because TransportAddress
is not
serializable in Elasticsearch 5.x.
List<E> httpHosts
RestClientFactory restClientFactory
List<E> httpHosts
RestClientFactory restClientFactory
AtomicReference<V> exceptionAtomicReference
com.google.api.core.ApiFutureCallback<V> failureHandler
AtomicInteger numPendingFutures
com.google.auth.Credentials credentials
SerializationSchema<T> serializationSchema
String projectName
String topicName
String hostAndPortForEmulator
boolean isRunning
PubSubDeserializationSchema<T> deserializationSchema
PubSubSubscriberFactory pubSubSubscriberFactory
com.google.auth.Credentials credentials
org.apache.flink.streaming.connectors.gcp.pubsub.PubSubSource.AcknowledgeOnCheckpointFactory acknowledgeOnCheckpointFactory
FlinkConnectorRateLimiter rateLimiter
int messagePerSecondRateLimit
private Object readResolve()
Properties properties
long pollTimeout
KafkaTopicsDescriptor topicsDescriptor
KafkaDeserializationSchema<T> deserializer
Map<K,V> subscribedPartitionsToStartOffsets
SerializedValue<T> watermarkStrategy
boolean enableCommitOnCheckpoints
boolean filterRestoredPartitionsWithCurrentTopicsDescriptor
OffsetCommitMode offsetCommitMode
FlinkKafkaConsumerBase.open(Configuration)
since it depends on whether or not checkpointing
is enabled for the job.long discoveryIntervalMillis
StartupMode startupMode
StartupMode.GROUP_OFFSETS
).Map<K,V> specificStartupOffsets
StartupMode.SPECIFIC_OFFSETS
.Long startupOffsetsTimestamp
StartupMode.TIMESTAMP
.org.apache.commons.collections.map.LinkedMap pendingOffsetsToCommit
boolean running
boolean useMetrics
true
, offset metrics
(e.g. current offset, committed offset) and Kafka-shipped metrics will be registered.FlinkKafkaErrorCode errorCode
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> keyedSchema
KafkaSerializationSchema<T> kafkaSchema
ProducerRecords
.FlinkKafkaPartitioner<T> flinkKafkaPartitioner
Map<K,V> topicPartitionsMap
int kafkaProducersPoolSize
BlockingDeque<E> availableTransactionalIds
boolean writeTimestampToKafka
boolean logFailuresOnly
FlinkKafkaProducer.Semantic semantic
AtomicLong pendingRecords
Map<K,V> previouslyCreatedMetrics
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> schema
FlinkKafkaPartitioner<T> flinkKafkaPartitioner
Map<K,V> topicPartitionsMap
boolean logFailuresOnly
boolean flushOnCheckpoint
SerializableObject pendingRecordsLock
long pendingRecords
DeserializationSchema<T> deserializationSchema
FlinkKafkaPartitioner<T> partitioner
SerializationSchema<T> serializationSchema
String topic
boolean writeTimestamp
int[] partitions
int parallelInstanceId
int numParallelInstances
TypeSerializer<T> typeSerializer
String topic
int partition
int cachedHash
int leaderId
int leaderPort
String leaderHost
KafkaTopicPartition topicPartition
int cachedHash
SerializationSchema<T> serializationSchema
int parallelInstanceId
TypeSerializer<T> typeSerializer
int producerParallelism
FlinkKafkaShuffleProducer.KafkaSerializer<IN> kafkaSerializer
KeySelector<IN,KEY> keySelector
int numberOfPartitions
TypeSerializer<T> serializer
int indexOfThisSubtask
List<E> streams
Properties configProps
KinesisDeserializationSchema<T> deserializer
KinesisShardAssigner shardAssigner
AssignerWithPeriodicWatermarks<T> periodicWatermarkAssigner
WatermarkTracker watermarkTracker
boolean running
Properties configProps
boolean failOnError
int queueLimit
String defaultStream
String defaultPartition
KinesisSerializationSchema<T> schema
KinesisPartitioner<T> customPartitioner
String streamName
com.amazonaws.services.kinesis.model.Shard shard
int cachedHash
String sequenceNumber
long subSequenceNumber
int cachedHash
String streamName
String shardId
String parentShardId
String adjacentParentShardId
String startingHashKey
String endingHashKey
String startingSequenceNumber
String endingSequenceNumber
DeserializationSchema<T> deserializationSchema
String delimiter
List<E> fieldNames
Map<K,V> dynamicFieldGetters
RowDataFieldsKinesisPartitioner.fieldNames
from an input record.StringBuilder keyBuffer
Map<K,V> staticFields
RowDataFieldsKinesisPartitioner.dynamicFieldGetters
entries.int keyBufferStaticPrefixLength
RowDataFieldsKinesisPartitioner.keyBuffer
(derived from the values in RowDataFieldsKinesisPartitioner.staticFields
).int fieldNamesStaticPrefixLength
RowDataFieldsKinesisPartitioner.fieldNames
for which
static field values are present in RowDataFieldsKinesisPartitioner.staticFields
.DeserializationSchema<T> physicalDeserializer
DeserializationSchema
to deserialize the physical part of the row.TypeInformation<T> producedTypeInfo
RowData
records (physical data with appended metadata].List<E> requestedMetadataFields
RowData
in the produced records.GlobalAggregateManager aggregateManager
String aggregateName
org.apache.flink.streaming.connectors.kinesis.util.JobManagerWatermarkTracker.WatermarkAggregateFunction aggregateFunction
long logAccumulatorIntervalMillis
long updateTimeoutCount
long watermark
long updateTimeoutCount
long watermark
String id
long updateTimeoutMillis
String subtaskId
org.apache.nifi.remote.client.SiteToSiteClient client
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
NiFiDataPacketBuilder<T> builder
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
long waitTimeMs
boolean isRunning
String queueName
RMQConnectionConfig rmqConnectionConfig
SerializationSchema<T> schema
boolean logFailuresOnly
RMQSinkPublishOptions<IN> publishOptions
SerializableReturnListener returnListener
RMQConnectionConfig rmqConnectionConfig
String queueName
boolean usesCorrelationId
RMQDeserializationSchema<T> deliveryDeserializer
String host
Integer port
String virtualHost
String username
String password
String uri
Integer networkRecoveryInterval
Boolean automaticRecovery
Boolean topologyRecovery
Integer connectionTimeout
Integer requestedChannelMax
Integer requestedFrameMax
Integer requestedHeartbeat
Integer prefetchCount
long deliveryTimeout
Properties properties
TwitterSource.EndpointInitializer initializer
Random rnd
Random rnd
int counter
int counter
double errorProbability
int delayPerRecordMillis
boolean running
InetAddress hostIp
int port
TypeSerializer<T> serializer
long currentTimestamp
String id
CheckpointCommitter committer
TypeSerializer<T> serializer
Set<E> pendingCheckpoints
WatermarkStrategy<T> watermarkStrategy
boolean emitProgressiveWatermarks
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
Class<T> typeClazz
SerializableSupplier<T> serializerSerializableSupplier
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
String previousSinkStateName
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
GlobalCommitter<CommT,GlobalCommT> globalCommitter
List<E> recoveredGlobalCommittables
boolean endOfInput
Sink<InputT,CommT,WriterStateT,GlobalCommT> sink
AssignerWithPeriodicWatermarks<T> wms
AssignerWithPunctuatedWatermarks<T> wms
WindowAssigner<T,W extends Window> windowAssigner
KeySelector<IN,KEY> keySelector
Trigger<T,W extends Window> trigger
StateDescriptor<S extends State,T> windowStateDescriptor
TypeSerializer<T> keySerializer
TypeSerializer<T> windowSerializer
long allowedLateness
window.maxTimestamp
+ allowedLateness
landmark.
OutputTag<T> lateDataOutputTag
OutputTag
to use for late arriving events. Elements for which window.maxTimestamp + allowedLateness
is smaller than the current watermark will be emitted
to this.Window window
WindowOperator.AbstractPerWindowStateStore windowState
AggregateFunction<IN,ACC,OUT> aggFunction
AggregateFunction<IN,ACC,OUT> aggFunction
InternalProcessWindowContext<IN,OUT,KEY,W extends Window> ctx
Window window
InternalWindowFunction.InternalWindowContext internalContext
Partitioner<K> partitioner
KeySelector<IN,KEY> keySelector
KeySelector<IN,KEY> keySelector
int maxParallelism
int nextChannelToSendTo
int nextChannelToSendTo
Random random
int numberOfChannels
TypeSerializer<T> typeSerializer
long numProcessedRecordsFailureThreshold
long numCompleteCheckpointsFailureThreshold
int maxNumFailures
long numProcessedRecords
long numCompleteCheckpoints
org.apache.flink.streaming.tests.SemanticsCheckMapper.ValidatorFunction validator
int payloadLength
int totalKeySpaceSize
long eventTimeClockProgressPerEvent
long maxOutOfOrder
long sleepTime
long sleepAfterElements
long monotonousEventTime
boolean running
TypeInformation<T> originalTypeInformation
int slideFactor
boolean failingTask
boolean killedJvm
int jvmPid
String taskNameWithSubtask
String allocationId
MapFunction<T,O> mapFunction
MapFunction<T,O> mapFunction
List<E> artificialStateBuilders
long eventTime
List<E> stringList
String strPayload
ComplexPayload.InnerPayLoad innerPayLoad
long sequenceNumber
AtomicReference<V> currentOwnerThread
ListStateDescriptor<T> listStateDescriptor
JoinFunction<IN1,IN2,OUT> keyedStateGenerator
JoinFunction<IN1,IN2,OUT> operatorStateGenerator
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
JoinFunction<IN1,IN2,OUT> stateValueGenerator
String stateName
ValueStateDescriptor<T> valueStateDescriptor
JoinFunction<IN1,IN2,OUT> stateValueGenerator
String emailId
Object value
long timestamp
Iterable<T> elements
boolean running
java.util.function.BooleanSupplier couldExit
long waitTimeOut
int[] fields
TypeComparator<T> comparator
int keyLength
Object[] keyArray
TypeComparator<T> comparator
Object[] keyArray
boolean includeMetadata
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
DataInputDeserializer inputDeserializer
TypeInformation<T> fieldType
private Object readResolve()
private Object readResolve()
String stateId
Field field
ListViewTypeInfo<T> listViewTypeInfo
private Object readResolve()
String stateId
Field field
MapViewTypeInfo<K,V> mapViewTypeInfo
private Object readResolve()
String name
org.apache.flink.table.expressions.WindowProperty property
private Object readResolve()
Object[] arrayKind
ArrayObjectArrayConverter<E> elementsConverter
Class<T> elementClass
int elementSize
BinaryArrayWriter.NullSetter writerNullSetter
BinaryWriter.ValueSetter writerValueSetter
org.apache.flink.table.data.conversion.ArrayObjectArrayConverter.GenericToJavaArrayConverter<E> genericToJavaArrayConverter
boolean hasInternalElements
ArrayData.ElementGetter elementGetter
DataStructureConverter<I,E> elementConverter
int precision
int scale
ArrayObjectArrayConverter<E> keyConverter
ArrayObjectArrayConverter<E> valueConverter
boolean hasInternalEntries
TypeSerializer<T> serializer
TypeSerializer<T> serializer
DataStructureConverter<I,E>[] fieldConverters
RowData.FieldGetter[] fieldGetters
DataStructureConverter<I,E>[] fieldConverters
RowData.FieldGetter[] fieldGetters
String generatedName
String generatedCode
org.apache.flink.table.data.conversion.YearMonthIntervalPeriodConverter.PeriodConstructor periodConstructor
DataFormatConverters.DataFormatConverter<Internal,External>[] converters
int precision
int scale
TupleTypeInfoBase<T> t
TupleSerializerBase<T> serializer
int precision
int scale
TypeSerializer<T> serializer
int precision
int precision
int precision
LogicalType keyType
LogicalType valueType
DataFormatConverters.DataFormatConverter<Internal,External> keyConverter
DataFormatConverters.DataFormatConverter<Internal,External> valueConverter
ArrayData.ElementGetter keyGetter
ArrayData.ElementGetter valueGetter
int keyElementSize
int valueElementSize
Class<T> keyComponentClass
Class<T> valueComponentClass
TypeSerializer<T> keySer
TypeSerializer<T> valueSer
boolean isKeyValueIndentity
Class<T> componentClass
LogicalType elementType
ArrayData.ElementGetter elementGetter
DataFormatConverters.DataFormatConverter<Internal,External> elementConverter
int elementSize
TypeSerializer<T> eleSer
boolean isEleIndentity
PojoTypeInfo<T> t
PojoField[] fields
int arity
int precision
int precision
int numRows
ColumnVector[] columns
boolean[] isNull
HeapIntVector dictionaryIds
boolean[] vector
int[] start
int[] length
byte[] buffer
int elementsAppended
int capacity
byte[] vector
double[] vector
float[] vector
int[] vector
long[] vector
short[] vector
long[] milliseconds
int[] nanoOfMilliseconds
boolean noNulls
Dictionary dictionary
TypeSerializer<T> listSerializer
TypeInformation<T> elementType
boolean nullSerializer
TypeSerializer<T> mapSerializer
TypeInformation<T> keyType
TypeInformation<T> valueType
boolean nullAware
boolean nullSerializer
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
List<E> parsingTypes
DynamicTableSource.DataStructureConverter converter
TypeInformation<T> producedTypeInfo
String columnDelimiter
String hostname
int port
byte byteDelimiter
DeserializationSchema<T> deserializer
boolean isRunning
Socket currentSocket
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
ImperativeAggregateFunction<T,ACC> aggregateFunction
TypeInformation<T> resultTypeInfo
TypeInformation<T> accTypeInfo
scala.collection.Seq<A> args
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
String name
scala.collection.Seq<A> extraNames
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
private Object readResolve()
scala.collection.Seq<A> elements
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression array
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression y
org.apache.flink.table.expressions.PlannerExpression x
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression expr
org.apache.flink.table.expressions.PlannerExpression lowerBound
org.apache.flink.table.expressions.PlannerExpression upperBound
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression container
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
scala.collection.Seq<A> strings
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression separator
scala.collection.Seq<A> strings
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression timestamp
org.apache.flink.table.expressions.PlannerExpression format
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlBinaryOperator sqlOperator
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression timeIntervalUnit
org.apache.flink.table.expressions.PlannerExpression temporal
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
Object key
scala.Option<A> fieldIndex
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression condition
org.apache.flink.table.expressions.PlannerExpression ifTrue
org.apache.flink.table.expressions.PlannerExpression ifFalse
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression expression
scala.collection.Seq<A> elements
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression container
org.apache.flink.table.expressions.PlannerExpression key
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression pattern
private Object readResolve()
Object value
TypeInformation<T> resultType
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression base
org.apache.flink.table.expressions.PlannerExpression antilogarithm
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression text
org.apache.flink.table.expressions.PlannerExpression len
org.apache.flink.table.expressions.PlannerExpression pad
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
scala.collection.Seq<A> elements
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlBinaryOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlFunction sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlBinaryOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression expr
org.apache.flink.table.expressions.PlannerExpression lowerBound
org.apache.flink.table.expressions.PlannerExpression upperBound
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlOperator sqlOperator
private Object readResolve()
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression agg
scala.collection.Seq<A> partitionBy
org.apache.flink.table.expressions.PlannerExpression orderBy
org.apache.flink.table.expressions.PlannerExpression preceding
org.apache.flink.table.expressions.PlannerExpression following
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression replacement
org.apache.flink.table.expressions.PlannerExpression starting
org.apache.flink.table.expressions.PlannerExpression position
private Object readResolve()
private Object readResolve()
String key
private Object readResolve()
String name
TypeInformation<T> resultType
private Object readResolve()
ScalarFunction scalarFunction
scala.collection.Seq<A> parameters
scala.Option<A> foundSignature
private Object readResolve()
String functionName
TableFunction<T> tableFunction
scala.collection.Seq<A> parameters
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue YEAR
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue YEAR_TO_MONTH
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue QUARTER
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MONTH
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue WEEK
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue DAY
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_HOUR
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_MINUTE
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_SECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue HOUR
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue HOUR_TO_MINUTE
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue HOUR_TO_SECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MINUTE
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MINUTE_TO_SECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue SECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue YEAR
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MONTH
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue DAY
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue HOUR
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MINUTE
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue SECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue QUARTER
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue WEEK
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MILLISECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue MICROSECOND
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue BOTH
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue LEADING
org.apache.flink.table.expressions.PlannerSymbols.PlannerSymbolValue TRAILING
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
org.apache.calcite.sql.SqlBinaryOperator sqlOperator
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression needle
org.apache.flink.table.expressions.PlannerExpression haystack
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression seed
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression seed
org.apache.flink.table.expressions.PlannerExpression bound
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression regex
org.apache.flink.table.expressions.PlannerExpression extractIndex
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression regex
org.apache.flink.table.expressions.PlannerExpression replacement
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression n
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression search
org.apache.flink.table.expressions.PlannerExpression replacement
private Object readResolve()
org.apache.calcite.rex.RexNode rexNode
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression left
org.apache.flink.table.expressions.PlannerExpression right
private Object readResolve()
scala.collection.Seq<A> elements
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression text
org.apache.flink.table.expressions.PlannerExpression len
org.apache.flink.table.expressions.PlannerExpression pad
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
org.apache.flink.table.expressions.PlannerExpression hashLength
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression pattern
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression str
org.apache.flink.table.expressions.PlannerExpression begin
org.apache.flink.table.expressions.PlannerExpression length
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerSymbol symbol
private Object readResolve()
String name
QueryOperation tableOperation
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression timeIntervalUnit
org.apache.flink.table.expressions.PlannerExpression temporal
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression timeIntervalUnit
org.apache.flink.table.expressions.PlannerExpression temporal
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression leftTimePoint
org.apache.flink.table.expressions.PlannerExpression leftTemporal
org.apache.flink.table.expressions.PlannerExpression rightTimePoint
org.apache.flink.table.expressions.PlannerExpression rightTemporal
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression timePointUnit
org.apache.flink.table.expressions.PlannerExpression timePoint1
org.apache.flink.table.expressions.PlannerExpression timePoint2
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression trimMode
org.apache.flink.table.expressions.PlannerExpression trimString
org.apache.flink.table.expressions.PlannerExpression str
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression base
org.apache.flink.table.expressions.PlannerExpression num
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
boolean valid
boolean bitmap$0
private Object readResolve()
String name
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression agg
org.apache.flink.table.expressions.PlannerExpression alias
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
String name
scala.Option<A> tpe
private Object readResolve()
org.apache.flink.table.expressions.PlannerExpression child
private Object readResolve()
DataGenerator<T> generator
DataGeneratorMapper.SerializableFunction<A,B> mapper
int precision
int scale
double min
double max
DataGenerator<T>[] fieldGenerators
String[] fieldNames
String pattern
DeserializationSchema<T> deserializationSchema
String[] fieldNames
DataType[] fieldTypes
int[] projectFields
RowType projectedRowType
List<E> partitionKeys
String defaultPartValue
int[] toProjectedField
RowData.FieldGetter[] formatFieldGetters
Path path
PartitionFetcher<P> partitionFetcher
PartitionFetcher.Context<P> fetcherContext
PartitionReader<P,OUT> partitionReader
RowData.FieldGetter[] lookupFieldGetters
java.time.Duration reloadInterval
TypeSerializer<T> serializer
RowType rowType
FileSystemFactory fsFactory
TableMetaStoreFactory msFactory
boolean overwrite
Path tmpPath
String[] partitionColumns
boolean dynamicGrouped
LinkedHashMap<K,V> staticPartitions
PartitionComputer<T> computer
OutputFormatFactory<T> formatFactory
OutputFileConfig outputFileConfig
BulkWriter.Factory<T> factory
RowDataPartitionComputer computer
PartitionComputer<T> computer
boolean rollOnCheckpoint
long rollingFileSize
long rollingTimeInterval
BulkFormat<T,SplitT extends FileSourceSplit> format
long limit
String defaultPartValue
String[] partitionColumns
int[] partitionIndexes
LogicalType[] partitionTypes
RowData.FieldGetter[] partitionFieldGetters
int[] nonPartitionIndexes
LogicalType[] nonPartitionTypes
RowData.FieldGetter[] nonPartitionFieldGetters
SerializationSchema<T> serializationSchema
long bucketCheckInterval
StreamingFileSink.BucketsBuilder<IN,BucketID,T extends StreamingFileSink.BucketsBuilder<IN,BucketID,T>> bucketsBuilder
Configuration conf
Path locationPath
ObjectIdentifier tableIdentifier
List<E> partitionKeys
TableMetaStoreFactory metaStoreFactory
FileSystemFactory fsFactory
SupplierWithException<R,E extends Throwable> fsFactory
long targetFileSize
long checkpointId
int taskId
int numberOfTasks
long checkpointId
SupplierWithException<R,E extends Throwable> fsFactory
CompactReader.Factory<T> readerFactory
CompactWriter.Factory<T> writerFactory
LogicalType deserializedType
TypeInformation<T> producedTypeInfo
String charsetName
boolean isBigEndian
org.apache.flink.table.formats.raw.RawFormatDeserializationSchema.DeserializationRuntimeConverter converter
org.apache.flink.table.formats.raw.RawFormatDeserializationSchema.DataLengthValidator validator
LogicalType serializedType
org.apache.flink.table.formats.raw.RawFormatSerializationSchema.SerializationRuntimeConverter converter
String charsetName
boolean isBigEndian
ObjectIdentifier objectIdentifier
String functionName
RowTypeInfo resultType
String className
HiveFunctionWrapper<UDFType> hiveFunctionWrapper
boolean isUDAFBridgeRequired
Object[] constantArguments
DataType[] argTypes
HiveShim hiveShim
HiveShim hiveShim
HiveFunctionWrapper<UDFType> hiveFunctionWrapper
Object[] constantArguments
DataType[] argTypes
HiveShim hiveShim
HiveFunctionWrapper<UDFType> hiveFunctionWrapper
Object[] constantArguments
DataType[] argTypes
HiveShim hiveShim
HiveObjectConversion flinkToJavaConversion
HiveShim hiveShim
String name
byte[] serializedAggregateFunction
DataType[] inputTypes
DataType resultType
DataType accumulatorType
PythonFunctionKind pythonFunctionKind
boolean deterministic
PythonEnv pythonEnv
int filterArg
boolean distinct
PythonEnv.ExecType execType
PythonFunction pythonFunction
Object[] inputs
String name
byte[] serializedScalarFunction
TypeInformation<T>[] inputTypes
TypeInformation<T> resultType
PythonFunctionKind pythonFunctionKind
boolean deterministic
PythonEnv pythonEnv
String name
byte[] serializedScalarFunction
TypeInformation<T>[] inputTypes
RowTypeInfo resultType
PythonFunctionKind pythonFunctionKind
boolean deterministic
PythonEnv pythonEnv
TemporalTableFunction temporalTableFunction
org.apache.calcite.rex.RexNode timeAttribute
private Object readResolve()
String name
org.apache.flink.table.planner.expressions.WindowProperty property
private Object readResolve()
String name
org.apache.flink.table.planner.expressions.PlannerWindowProperty property
private Object readResolve()
String keyTerm
org.apache.calcite.rel.type.RelDataType dataType
LogicalType internalType
private Object readResolve()
String fieldTerm
org.apache.calcite.rel.type.RelDataType dataType
private Object readResolve()
ImperativeAggregateFunction<T,ACC> aggregateFunction
TypeInformation<T> resultTypeInfo
TypeInformation<T> accTypeInfo
scala.collection.Seq<A> args
private Object readResolve()
CallExpression resolvedCall
scala.collection.immutable.Nil$ children
private Object readResolve()
DataType resolvedDataType
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression array
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression container
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression timestamp
org.apache.flink.table.planner.expressions.PlannerExpression format
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression timeIntervalUnit
org.apache.flink.table.planner.expressions.PlannerExpression temporal
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression expression
scala.collection.Seq<A> elements
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression container
org.apache.flink.table.planner.expressions.PlannerExpression key
private Object readResolve()
Object value
TypeInformation<T> resultType
private Object readResolve()
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression agg
scala.collection.Seq<A> partitionBy
org.apache.flink.table.planner.expressions.PlannerExpression orderBy
org.apache.flink.table.planner.expressions.PlannerExpression preceding
org.apache.flink.table.planner.expressions.PlannerExpression following
private Object readResolve()
String name
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerWindowReference reference
private Object readResolve()
String name
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerWindowReference reference
private Object readResolve()
ScalarFunction scalarFunction
scala.collection.Seq<A> parameters
LogicalType[] signature
private Object readResolve()
String functionName
TableFunction<T> tableFunction
scala.collection.Seq<A> parameters
TypeInformation<T> resultType
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue YEAR
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue YEAR_TO_MONTH
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue QUARTER
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MONTH
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue WEEK
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue DAY
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_HOUR
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_MINUTE
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue DAY_TO_SECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue HOUR
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue HOUR_TO_MINUTE
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue HOUR_TO_SECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MINUTE
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MINUTE_TO_SECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue SECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue YEAR
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MONTH
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue DAY
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue HOUR
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MINUTE
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue SECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue QUARTER
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue WEEK
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MILLISECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue MICROSECOND
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue BOTH
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue LEADING
org.apache.flink.table.planner.expressions.PlannerSymbols.PlannerSymbolValue TRAILING
org.apache.flink.table.planner.expressions.PlannerWindowReference reference
private Object readResolve()
String name
scala.Option<A> tpe
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerWindowReference reference
private Object readResolve()
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
TypeInformation<T> resultType
boolean checkOverflow
private Object readResolve()
org.apache.calcite.rex.RexNode rexNode
private Object readResolve()
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerSymbol symbol
private Object readResolve()
String name
QueryOperation tableOperation
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression leftTimePoint
org.apache.flink.table.planner.expressions.PlannerExpression leftTemporal
org.apache.flink.table.planner.expressions.PlannerExpression rightTimePoint
org.apache.flink.table.planner.expressions.PlannerExpression rightTemporal
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression msg
TypeInformation<T> tp
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression timePointUnit
org.apache.flink.table.planner.expressions.PlannerExpression timePoint1
org.apache.flink.table.planner.expressions.PlannerExpression timePoint2
private Object readResolve()
private Object readResolve()
private Object readResolve()
String name
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression agg
org.apache.flink.table.planner.expressions.PlannerExpression alias
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
String name
scala.Option<A> tpe
private Object readResolve()
org.apache.flink.table.planner.expressions.PlannerExpression child
private Object readResolve()
UnresolvedReferenceExpression sum
UnresolvedReferenceExpression count
DecimalType type
UnresolvedReferenceExpression count1
UnresolvedReferenceExpression count
UnresolvedReferenceExpression sum
DecimalType decimalType
UnresolvedReferenceExpression sum
UnresolvedReferenceExpression count
DecimalType decimalType
int operandCount
boolean existDefaultValue
UnresolvedReferenceExpression value
DecimalType decimalType
TimestampType type
int operandCount
UnresolvedReferenceExpression acc
UnresolvedReferenceExpression accDelimiter
Expression delimiter
Expression operand
UnresolvedReferenceExpression max
DecimalType decimalType
TimestampType type
UnresolvedReferenceExpression min
DecimalType decimalType
TimestampType type
UnresolvedReferenceExpression currNumber
UnresolvedReferenceExpression sequence
UnresolvedReferenceExpression[] lastValues
LogicalType[] orderKeyTypes
UnresolvedReferenceExpression sequence
UnresolvedReferenceExpression value
UnresolvedReferenceExpression count
DecimalType type
TimestampType type
UnresolvedReferenceExpression sum0
DecimalType decimalType
UnresolvedReferenceExpression sum
DecimalType decimalType
UnresolvedReferenceExpression sum
UnresolvedReferenceExpression count
DecimalType decimalType
TypeInformation<T>[] fieldTypes
TemporalTableFunction temporalTableFunction
org.apache.calcite.rex.RexNode timeAttribute
private Object readResolve()
TypeInformationRawType<T> genericType
boolean nullable
StructuredType structuredType
org.apache.calcite.rel.type.RelDataTypeSystem typeSystem
org.apache.calcite.sql.type.BasicSqlType originalType
boolean nullable
boolean isEventTime
long interval
scala.Enumeration.Value mode
private Object readResolve()
org.apache.flink.table.planner.plan.trait.MiniBatchInterval NONE
org.apache.flink.table.planner.plan.trait.MiniBatchInterval NO_MINIBATCH
scala.Enumeration.Value ProcTime
scala.Enumeration.Value RowTime
scala.Enumeration.Value None
org.apache.calcite.rel.core.AggregateCall agg
UserDefinedFunction function
int aggIndex
int[] argIndexes
DataType[] externalArgTypes
DataType[] externalAccTypes
DataViewUtils.DataViewSpec[] viewSpecs
DataType externalResultType
boolean consumeRetraction
private Object readResolve()
org.apache.flink.table.planner.plan.utils.AggregateInfo[] aggInfos
scala.Option<A> indexOfCountStar
boolean countStarInserted
org.apache.flink.table.planner.plan.utils.DistinctInfo[] distinctInfos
private Object readResolve()
private Object readResolve()
int[] argIndexes
DataType keyType
DataType accType
boolean excludeAcc
scala.Option<A> dataViewSpec
boolean consumeRetraction
scala.collection.mutable.ArrayBuffer<A> filterArgs
scala.collection.mutable.ArrayBuffer<A> aggIndexes
private Object readResolve()
boolean isEventTime
boolean isLeftInput
int idx
private Object readResolve()
boolean isEventTime
boolean leftInputOnLeftSide
int leftTimeIdx
int rightTimeIdx
org.apache.calcite.rex.RexCall pred
private Object readResolve()
long bound
boolean isLeftLower
private Object readResolve()
boolean isEventTime
long leftLowerBound
long leftUpperBound
int leftTimeIdx
int rightTimeIdx
private Object readResolve()
LogicalType dataType
RexLiteral literal
private Object readResolve()
int index
private Object readResolve()
long lower
long upper
private Object readResolve()
int inputFieldIndex
private Object readResolve()
boolean rankOnLeftSide
org.apache.calcite.rex.RexCall pred
private Object readResolve()
long lower
private Object readResolve()
long upper
private Object readResolve()
private Object readResolve()
private Object readResolve()
int[] primaryKeys
private Object readResolve()
String id
TypeSerializer<T> typeSerializer
SerializedListAccumulator<T> accumulator
TypeSerializer<T> elementSerializer
boolean containsNullKey
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
DataType dataType
byte[][] arrowData
Object input
ResultFuture<OUT> resultFuture
long minRetentionTime
long maxRetentionTime
boolean stateCleaningEnabled
ValueState<T> cleanupTimeState
ArrayData.ElementGetter elementGetter
ArrayData.ElementGetter keyGetter
ArrayData.ElementGetter valueGetter
PythonFunctionInfo[] scalarFunctions
ScalarFunction
s to be executed.int[] forwardedFields
PythonConfig config
int[] userDefinedFunctionInputOffsets
RowType inputType
RowType outputType
Map<K,V> jobOptions
PythonFunctionInfo tableFunction
TableFunction
to be executed.org.apache.calcite.rel.core.JoinRelType joinType
int maxArrowBatchSize
String name
String code
GenericInputFormat<OT> format
String name
String code
GenericInputFormat<OT> format
InternalTypeInfo<T> keyRowType
GeneratedProjection generatedProjection
InternalTypeInfo<T> returnType
long currentWatermark
InternalTimerService
.GeneratedClass<T> generatedClass
CoProcessFunction<IN1,IN2,OUT> flatMapper
long watermarkDelay
scala.Function1<T1,R> emitter
private Object readResolve()
ProcessFunction<I,O> function
long watermarkDelay
scala.Function1<T1,R> emitter
private Object readResolve()
boolean closed
GeneratedAggsHandleFunction genAggsHandler
GeneratedRecordEqualiser genRecordEqualiser
LogicalType[] accTypes
RecordCounter recordCounter
boolean generateUpdateBefore
long stateRetentionTime
GeneratedTableAggsHandleFunction genAggsHandler
LogicalType[] accTypes
RecordCounter recordCounter
boolean generateUpdateBefore
long stateRetentionTime
GeneratedAggsHandleFunction genLocalAggsHandler
GeneratedAggsHandleFunction genGlobalAggsHandler
GeneratedRecordEqualiser genRecordEqualiser
LogicalType[] accTypes
RecordCounter recordCounter
boolean generateUpdateBefore
long stateRetentionTime
GeneratedAggsHandleFunction genAggsHandler
GeneratedRecordEqualiser genRecordEqualiser
LogicalType[] accTypes
RowType inputType
RecordCounter recordCounter
boolean generateUpdateBefore
long stateRetentionTime
GeneratedAggsHandleFunction genPartialAggsHandler
GeneratedAggsHandleFunction genFinalAggsHandler
KeySelector<IN,KEY> finalKeySelector
long stateRetentionTime
GeneratedAggsHandleFunction genAggsHandler
BundleTrigger<T> bundleTrigger
MapBundleFunction<K,V,IN,OUT> function
KeySelector<IN,KEY> keySelector
long maxCount
long maxCount
boolean generateUpdateBefore
boolean generateInsert
boolean inputIsInsertOnly
TypeSerializer<T> serializer
TypeSerializer<T> serializer
boolean generateUpdateBefore
boolean generateInsert
boolean inputInsertOnly
boolean generateUpdateBefore
boolean generateInsert
int rowtimeIndex
boolean keepLastRow
TypeSerializer<T> serializer
boolean generateUpdateBefore
boolean generateInsert
int rowtimeIndex
boolean keepLastRow
org.apache.flink.table.runtime.operators.join.HashJoinOperator.HashJoinParameter parameter
boolean reverseJoinFunction
HashJoinType type
java.util.function.Consumer<T> emitter
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
int leftArity
int rightArity
double externalBufferMemRatio
FlinkJoinType type
boolean leftIsSmaller
boolean[] filterNulls
GeneratedJoinCondition condFuncCode
GeneratedProjection projectionCode1
GeneratedProjection projectionCode2
GeneratedNormalizedKeyComputer computer1
GeneratedRecordComparator comparator1
GeneratedNormalizedKeyComputer computer2
GeneratedRecordComparator comparator2
GeneratedRecordComparator genKeyComparator
int leftTimeIdx
int rightTimeIdx
GeneratedFunction<F extends Function> generatedFetcher
GeneratedResultFuture<T extends ResultFuture<?>> generatedResultFuture
boolean isLeftOuterJoin
int asyncBufferCapacity
TypeInformation<T> fetcherReturnType
InternalTypeInfo<T> rightRowTypeInfo
GeneratedFunction<F extends Function> generatedCalc
InternalTypeInfo<T> rightRowTypeInfo
GeneratedFunction<F extends Function> generatedFetcher
GeneratedCollector<C extends Collector<?>> generatedCollector
boolean isLeftOuterJoin
int tableFieldsCount
GeneratedFunction<F extends Function> generatedCalc
GeneratedJoinCondition generatedJoinCondition
InternalTypeInfo<T> leftType
InternalTypeInfo<T> rightType
JoinInputSideSpec leftInputSideSpec
JoinInputSideSpec rightInputSideSpec
int[] nullFilterKeys
boolean nullSafe
boolean filterAllNulls
long stateRetentionTime
boolean leftIsOuter
boolean rightIsOuter
boolean isAntiJoin
boolean inputSideHasUniqueKey
boolean joinKeyContainsUniqueKey
InternalTypeInfo<T> uniqueKeyType
KeySelector<IN,KEY> uniqueKeySelector
long minRetentionTime
long maxRetentionTime
boolean stateCleaningEnabled
boolean isLeftOuterJoin
InternalTypeInfo<T> rightType
GeneratedJoinCondition generatedJoinCondition
boolean isLeftOuterJoin
InternalTypeInfo<T> leftType
InternalTypeInfo<T> rightType
GeneratedJoinCondition generatedJoinCondition
int leftTimeAttribute
int rightTimeAttribute
org.apache.flink.table.runtime.operators.join.temporal.TemporalRowTimeJoinOperator.RowtimeComparator rightRowtimeComparator
GeneratedFunction<F extends Function> generatedFunction
GeneratedFunction<F extends Function> generatedFunction
GeneratedRecordComparator generatedComparator
int rowtimeIdx
int precision
InputSelectionHandler inputSelectionHandler
List<E> inputSpecs
List<E> headWrappers
TableOperatorWrapper<OP extends StreamOperator<RowData>> tailWrapper
List<E> inputSpecs
Map<K,V> inputSpecMap
List<E> headWrappers
TableOperatorWrapper<OP extends StreamOperator<RowData>> tailWrapper
Deque<E> topologicalOrderingOperators
StreamOperatorFactory<OUT> factory
String operatorName
List<E> allInputTypes
NOTE:The inputs of an operator may not all be in the multiple-input operator, e.g. The multiple-input operator contains A and J, and A is one of the input of J, and another input of J is not in the multiple-input operator.
------- \ J -- / -- A --
For this example, `allInputTypes` contains two input types.
TypeInformation<T> outputType
double managedMemoryFraction
List<E> inputEdges
List<E> outputEdges
boolean closed
int endedInputCount
TableOperatorWrapper<OP extends StreamOperator<RowData>> source
TableOperatorWrapper<OP extends StreamOperator<RowData>> target
int inputId
int multipleInputId
Input
in MultipleInputStreamOperatorBase.getInputs()
.int readOrder
TableOperatorWrapper<OP extends StreamOperator<RowData>> output
Input
.int outputOpInputId
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
LogicalType[] inputFieldTypes
int rowTimeIdx
OverWindowFrame[] overWindowFrames
GeneratedRecordComparator genComparator
boolean isRowAllInFixedPart
RecordComparator partitionComparator
RowData lastInput
JoinedRowData[] joinedRows
StreamRecordCollector<T> collector
AbstractRowDataSerializer<T extends RowData> serializer
ResettableExternalBuffer currentData
GeneratedAggsHandleFunction[] aggsHandlers
GeneratedRecordComparator genComparator
boolean[] resetAccumulators
RecordComparator partitionComparator
RowData lastInput
AggsHandleFunction[] processors
JoinedRowData[] joinedRows
StreamRecordCollector<T> collector
AbstractRowDataSerializer<T extends RowData> serializer
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
LogicalType[] inputFieldTypes
long precedingTimeBoundary
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
LogicalType[] inputFieldTypes
long precedingOffset
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
LogicalType[] inputFieldTypes
long precedingOffset
int rowTimeIdx
GeneratedAggsHandleFunction genAggsHandler
LogicalType[] accTypes
LogicalType[] inputFieldTypes
long precedingOffset
int rowTimeIdx
GeneratedAggsHandleFunction aggsHandleFunction
AggsHandleFunction processor
GeneratedAggsHandleFunction aggsHandleFunction
Long offset
OffsetOverFrame.CalcOffsetFunc calcOffsetFunc
AggsHandleFunction processor
ResettableExternalBuffer.BufferIterator inputIterator
long inputIndex
ResettableExternalBuffer externalBuffer
long currentBufferLength
GeneratedRecordComparator lboundComparator
GeneratedRecordComparator rboundComparator
RecordComparator lbound
RecordComparator rbound
GeneratedRecordComparator boundComparator
RecordComparator lbound
GeneratedRecordComparator boundComparator
RecordComparator rbound
long leftBound
long rightBound
int inputRightIndex
int inputLeftIndex
long leftBound
long rightBound
long inputRightIndex
RowType inputType
RowType valueType
GeneratedAggsHandleFunction aggsHandleFunction
GeneratedAggsHandleFunction aggsHandleFunction
RowType valueType
AggsHandleFunction processor
RowData accValue
ResettableExternalBuffer input
RowDataSerializer valueSer
int inputIndex
GeneratedAggsHandleFunction aggsHandleFunction
RowType valueType
AggsHandleFunction processor
RowData accValue
RowDataSerializer valueSer
GeneratedAggsHandleFunction aggsHandleFunction
AggsHandleFunction processor
RowData accValue
ResettableExternalBuffer.BufferIterator inputIterator
BinaryRowData nextRow
RowType inputType
RowType outputType
int[] userDefinedFunctionInputOffsets
Map<K,V> jobOptions
RowType inputType
RowType outputType
Map<K,V> jobOptions
PythonAggregateFunctionInfo[] aggregateFunctions
DataViewUtils.DataViewSpec[][] dataViewSpecs
int[] grouping
int indexOfCountStar
boolean countStarInserted
boolean generateUpdateBefore
long minRetentionTime
long maxRetentionTime
int stateCacheSize
int mapStateReadCacheSize
int mapStateWriteCacheSize
boolean stateCleaningEnabled
PythonFunctionInfo[] pandasAggFunctions
AggregateFunction
s to be executed.int[] groupingSet
int[] namedProperties
int inputTimeFieldIndex
int maxLimitSize
long windowSize
long slideSize
long[] lowerBoundary
long[] upperBoundary
boolean[] isRangeWindows
int[] aggWindowIndex
int inputTimeFieldIndex
boolean asc
long minRetentionTime
long maxRetentionTime
boolean stateCleaningEnabled
int inputTimeFieldIndex
long lowerBoundary
int[] namedProperties
int inputTimeFieldIndex
WindowAssigner<W extends Window> windowAssigner
WindowAssigner
assigns zero or more Windows
to an element.Trigger<W extends Window> trigger
Trigger
determines when a pane of a window should be evaluated to emit the results
for that part of the window.long allowedLateness
window.maxTimestamp
+ allowedLateness
landmark.
PythonFunctionInfo[] scalarFunctions
ScalarFunction
s to be executed.int[] forwardedFields
PythonFunctionInfo tableFunction
TableFunction
to be executed.org.apache.calcite.rel.core.JoinRelType joinType
GeneratedRecordComparator generatedSortKeyComparator
Comparator<T> sortKeyComparator
boolean generateUpdateBefore
boolean outputRankNumber
InternalTypeInfo<T> inputRowType
KeySelector<IN,KEY> sortKeySelector
KeyContext keyContext
boolean isConstantRankEnd
long rankStart
int rankEndIndex
long rankEnd
ValueState<T> rankEndState
Counter invalidCounter
JoinedRowData outputRow
long hitCount
long requestCount
InternalTypeInfo<T> sortKeyType
TypeSerializer<T> inputRowSer
long cacheSize
GeneratedRecordComparator generatedRecordComparator
int[] compareKeyPositions
LogicalType[] compareKeyTypes
boolean[] compareOrders
boolean[] nullsIsLast
long rankStart
long rankEnd
long rankStart
InternalTypeInfo<T> sortKeyType
boolean lenient
GeneratedRecordEqualiser generatedEqualiser
RecordEqualiser equaliser
ComparableRecordComparator serializableComparator
InternalTypeInfo<T> rowKeyType
long cacheSize
TypeSerializer<T> inputRowSer
KeySelector<IN,KEY> rowKeySelector
int rankEndIndex
ExecutionConfigOptions.NotNullEnforcer notNullEnforcer
int[] notNullFieldIndices
boolean notNullCheck
String[] allFieldNames
int rowtimeFieldIndex
SinkNotNullEnforcer enforcer
long currentWatermark
InternalTimerService
.boolean isGlobal
long limitStart
long limitEnd
InternalTypeInfo<T> inputRowType
GeneratedRecordComparator gComparator
GeneratedRecordComparator partitionByGenComp
GeneratedRecordComparator orderByGenComp
long rankStart
long rankEnd
boolean outputRankFunColumn
InternalTypeInfo<T> inputRowType
int rowTimeIdx
GeneratedRecordComparator gComparator
boolean isGlobal
long limitStart
long limitEnd
GeneratedRecordComparator genComparator
GeneratedNormalizedKeyComputer gComputer
GeneratedRecordComparator gComparator
InternalTypeInfo<T> inputRowType
GeneratedRecordComparator gComparator
GeneratedInput<F extends InputFormat<?,?>> generatedInput
InternalTypeInfo<T> returnType
GenericInputFormat<OT> format
NamespaceAggsHandleFunction<N> aggWindowAggregator
GeneratedNamespaceAggsHandleFunction<N> generatedAggWindowAggregator
RecordEqualiser equaliser
GeneratedRecordEqualiser generatedEqualiser
NamespaceTableAggsHandleFunction<N> tableAggWindowAggregator
GeneratedNamespaceTableAggsHandleFunction<N> generatedTableAggWindowAggregator
WindowAssigner<W extends Window> windowAssigner
Trigger<W extends Window> trigger
TypeSerializer<T> windowSerializer
LogicalType[] inputFieldTypes
LogicalType[] accumulatorTypes
LogicalType[] aggResultTypes
LogicalType[] windowPropertyTypes
boolean produceUpdates
int rowtimeIndex
long allowedLateness
window.maxTimestamp
+ allowedLateness
landmark.
NamespaceAggsHandleFunctionBase<N> windowAggregator
long windowSize
long windowSlide
long size
long maxSize
long step
long offset
boolean isEventTime
long sessionGap
boolean isEventTime
long size
long slide
long offset
long paneSize
int numPanesPerWindow
boolean isEventTime
long size
long offset
boolean isEventTime
WindowAssigner<W extends Window> windowAssigner
NamespaceAggsHandleFunctionBase<N> windowAggregator
long allowedLateness
InternalWindowProcessFunction.Context<K,W extends Window> ctx
MergingWindowAssigner<W extends Window> windowAssigner
TypeSerializer<T> windowSerializer
List<E> reuseActualWindows
PanedWindowAssigner<W extends Window> windowAssigner
long countElems
ReducingStateDescriptor<T> countStateDesc
Trigger.TriggerContext ctx
Trigger<W extends Window> earlyTrigger
Trigger.TriggerContext ctx
Trigger.TriggerContext ctx
Trigger<W extends Window> earlyTrigger
Trigger.TriggerContext ctx
long interval
ReducingStateDescriptor<T> nextFiringStateDesc
long delay
int rowtimeIndex
long intervalMs
long minibatchInterval
int rowtimeFieldIndex
long idleTimeout
WatermarkGenerator watermarkGenerator
int rowtimeFieldIndex
long idleTimeout
GeneratedWatermarkGenerator generatedWatermarkGenerator
GeneratedHashFunction genHashFunc
String[] hashFieldNames
TypeComparator<T> rowComp
TypeSerializer<T> rowSerializer
private Object readResolve()
RowTypeInfo rowType
private Object readResolve()
LogicalType eleType
TypeSerializer<T> eleSer
ArrayData.ElementGetter elementGetter
int precision
int scale
int numFields
int fixedLengthPartSize
int precision
int scale
int precision
int scale
private void readObject(ObjectInputStream serialized) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
DataType dataType
TypeSerializer<T> internalSerializer
boolean isReuseEnabled
DataType dataType
TypeSerializer<T> typeSerializer
LogicalType type
Class<T> typeClass
TypeSerializer<T> typeSerializer
int precision
int precision
int precision
LogicalType keyType
LogicalType valueType
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
ArrayData.ElementGetter keyGetter
ArrayData.ElementGetter valueGetter
TypeSerializer<T> serializer
BinaryRowDataSerializer binarySerializer
LogicalType[] types
TypeSerializer<T>[] fieldSerializers
RowData.FieldGetter[] fieldGetters
Comparator<T> comparator
Comparator<T> comparator
int precision
int precision
LogicalType elementType
TypeSerializer<T> elementTypeSerializer
ArrayData.ElementGetter elementGetter
int elementSize
BinaryArrayWriter.NullSetter nullSetter
int precision
int scale
LogicalType keyType
LogicalType valueType
TypeSerializer<T> keyTypeSerializer
TypeSerializer<T> valueTypeSerializer
BinaryWriter.ValueSetter keySetter
BinaryWriter.ValueSetter valueSetter
ArrayData.ElementGetter keyGetter
ArrayData.ElementGetter valueGetter
BinaryArrayWriter.NullSetter nullValueSetter
int keySize
int valueSize
LogicalType[] fieldTypes
TypeSerializer<T>[] fieldSerializers
RowData.FieldGetter[] fieldGetters
int precision
int cacheSize
LRUMap.RemovalListener<K,V> removalListener
String fieldDelim
String field
long maxTimestamp
long delay
long maxTimestamp
DataType elementDataType
LogicalType logicalType
Class<T> conversionClass
LogicalType elementType
int length
int length
DayTimeIntervalType.DayTimeResolution resolution
int dayPrecision
int fractionalPrecision
int precision
int scale
LogicalType sourceType
TypeInformation<T> typeInfo
TimestampKind kind
int precision
boolean isNullable
LogicalTypeRoot typeRoot
LogicalType keyType
LogicalType valueType
LogicalType elementType
Class<T> clazz
TypeSerializer<T> serializer
String name
LogicalType type
String description
List<E> attributes
boolean isInstantiable
StructuredType.StructuredComparision comparision
StructuredType superType
Class<T> implementationClass
String name
LogicalType type
String description
TimestampKind kind
int precision
int precision
TypeInformation<T> typeInfo
UnresolvedIdentifier unresolvedIdentifier
ObjectIdentifier objectIdentifier
boolean isFinal
String description
int length
int length
YearMonthIntervalType.YearMonthResolution resolution
int yearPrecision
TimestampKind kind
int precision
boolean isEventTime
boolean cancel
boolean value
byte value
char value
double value
float value
int value
int fieldNumber
long value
int fieldPos
int fieldNumber
org.apache.flink.types.Record.InternalDeSerializer serializer
byte[] binaryData
byte[] switchBuffer
int[] offsets
int[] lengths
Value[] readFields
Value[] writeFields
int binaryLen
int numFields
int firstModifiedPos
Serializable value
short value
char[] value
int len
int hashCode
long upperPart
long lowerPart
long to
long current
LongValue currentValue
long to
long current
private void readObject(ObjectInputStream stream) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream stream) throws IOException
IOException
Throwable failureCause
String id
TypeInformation<T> typeInfo
SerializableObject lock
int leaseCount
boolean closed
byte[] serializedData
String keyString
StringValue toTokenize
int pos
int limit
long seed
int status
int status
int exitCode
Copyright © 2014–2021 The Apache Software Foundation. All rights reserved.