Collator collator
String prefix
ExecutionMode executionMode
ExecutionConfig.ClosureCleanerLevel closureCleanerLevel
int parallelism
int maxParallelism
int numberOfExecutionRetries
boolean forceKryo
boolean disableGenericTypes
boolean enableAutoGeneratedUids
boolean objectReuse
boolean autoTypeRegistrationEnabled
boolean forceAvro
long autoWatermarkInterval
long latencyTrackingInterval
boolean isLatencyTrackingConfigured
long executionRetryDelay
RestartStrategies.RestartStrategyConfiguration restartStrategyConfiguration
long taskCancellationIntervalMillis
long taskCancellationTimeoutMillis
boolean useSnapshotCompression
ExecutionConfig.GlobalJobParameters globalJobParameters
LinkedHashMap<K,V> registeredTypesWithKryoSerializers
LinkedHashMap<K,V> registeredTypesWithKryoSerializerClasses
LinkedHashMap<K,V> defaultKryoSerializers
LinkedHashMap<K,V> defaultKryoSerializerClasses
LinkedHashSet<E> registeredKryoTypes
LinkedHashSet<E> registeredPojoTypes
com.esotericsoftware.kryo.Serializer<T> serializer
long count
double sum
double localValue
double max
double min
int localValue
int max
int min
long localValue
long max
long min
DoubleValue wrapper
double sum
long sum
TypeComparator<T> typeComparator
Object[][] boundaries
TypeComparator<T>[] flatComparators
Object[] keys
SerializableTimestampAssigner<T> assigner
long timestamp
boolean target
String sinkIdentifier
long blockSize
long readRecords
long blockSize
Path filePath
FileInputFormat.supportsMultiPaths()
and use FileInputFormat.getFilePaths()
and FileInputFormat.setFilePaths(Path...)
.Path[] filePaths
long minSplitSize
int numSplits
long openTimeout
boolean unsplittable
boolean enumerateNestedFiles
FilePathFilter filesFilter
Path outputFilePath
FileSystem.WriteMode writeMode
FileOutputFormat.OutputDirectoryMode outputDirectoryMode
boolean lineDelimiterIsLinebreak
Class<T>[] fieldTypes
boolean[] fieldIncluded
byte[] fieldDelim
String fieldDelimString
boolean lenient
boolean skipFirstLineAsHeader
boolean quotedStringParsing
byte quoteCharacter
byte[] commentPrefix
String commentPrefixString
int partitionNumber
InputFormat<OT,T extends InputSplit> replicatedIF
long globalRateBytesPerSecond
long localRateBytesPerSecond
RuntimeContext runtimeContext
org.apache.flink.shaded.guava18.com.google.common.util.concurrent.RateLimiter rateLimiter
Map<K,V> fieldMapping1
Map<K,V> fieldMapping2
FieldSet readFields1
FieldSet readFields2
private Object readResolve()
CPUResource cpuCores
MemorySize taskHeapMemory
MemorySize taskOffHeapMemory
MemorySize managedMemory
Map<K,V> extendedResources
long count
BulkIterationBase.TerminationCriterionAggregator aggregator
Object userCodeObject
String name
BigDecimal value
int restartAttempts
Time delayBetweenAttemptsInterval
TypeInformation<T> type
DeserializationSchema
.String charsetName
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeInformation<T> typeInfo
TypeInformationSerializationSchema.getProducedType()
.TypeSerializer<T> serializer
AggregateFunction<IN,ACC,OUT> aggFunction
ReduceFunction<T> reduceFunction
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
String name
AtomicReference<V> serializerAtomicReference
StateDescriptor.initializeSerializerUnlessSet(ExecutionConfig)
method is called.TypeInformation<T> typeInfo
String queryableStateName
StateTtlConfig ttlConfig
StateTtlConfig.UpdateType updateType
StateTtlConfig.StateVisibility stateVisibility
StateTtlConfig.TtlTimeCharacteristic ttlTimeCharacteristic
Time ttl
StateTtlConfig.CleanupStrategies cleanupStrategies
int cleanupSize
boolean runCleanupForEveryRecord
long queryTimeAfterNumEntries
TimeUnit unit
long size
Class<T> arrayClass
TypeInformation<T> componentInfo
TypeSerializer<T>[] fieldSerializers
CompositeSerializer.PrecomputedParameters precomputed
boolean immutableTargetType
boolean immutable
int length
boolean stateful
TypeComparator<T> comparator1
TypeComparator<T> comparator2
TypeComparator<T>[] comparators1
TypeComparator<T>[] comparators2
Object[] referenceKeyFields
Object[] candidateKeyFields
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
org.apache.flink.api.common.typeutils.SingleThreadAccessCheckingTypeSerializer.SingleThreadAccessChecker singleThreadAccessChecker
TypeSerializer<T> originalSerializer
byte[] actualBytes
Throwable originalError
boolean ascendingComparison
TypeComparator<T>[] comparators
boolean ascendingComparison
BooleanValue reference
BooleanValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
ByteValue reference
ByteValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
CharValue reference
CharValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
DoubleValue reference
DoubleValue tempReference
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> enumClass
Map<K,V> valueToOrdinal
Enum.ordinal()
. This allows us to maintain backwards compatibility for previous serialized
data in the case that the order of enum constants was changed or new constants were added.
On a fresh start with no reconfiguration, the ordinals would simply be identical to the enum constants actual ordinals. Ordinals may change after reconfiguration.
Enum<E extends Enum<E>>[] values
EnumSerializer.valueToOrdinal
map. Serves as a bidirectional map to have fast access from ordinal to
value. May be reordered after reconfiguration.boolean ascendingComparison
FloatValue reference
FloatValue tempReference
TypeComparator<T>[] comparators
Class<T> componentClass
TypeSerializer<T> componentSerializer
boolean ascendingComparison
IntValue reference
IntValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> elementSerializer
boolean ascendingComparison
LocalDateComparator[] comparators
boolean ascendingComparison
LocalDateComparator dateComparator
LocalTimeComparator timeComparator
LocalDateTimeComparator[] comparators
boolean ascendingComparison
LongValue reference
LongValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeComparator<T>[] comparators
boolean ascendingComparison
ShortValue reference
ShortValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
StringValue reference
StringValue tempReference
TypeComparator<T>[] comparators
TypeComparator<T>[] comparators
boolean ascending
BasicTypeComparator<T extends Comparable<T>> comparator
int subtaskId
String location
long from
long to
long count
long checksum
String id
long counter
long checksum
String id
TypeSerializer<T> serializer
SerializedListAccumulator<T> accumulator
String id
long counter
int count
TextOutputFormat.TextFormatter<IN> formatter
boolean withReplacement
int numSample
long seed
boolean withReplacement
int numSample
long seed
boolean withReplacement
double fraction
long seed
int[] fields
int[] fields
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapred.JobConf jobConf
org.apache.hadoop.mapred.OutputFormat<K,V> mapredOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapreduce.InputFormat<K,V> mapreduceInputFormat
Class<T> keyClass
Class<T> valueClass
Configuration configuration
boolean fetched
boolean hasNext
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration configuration
org.apache.hadoop.mapreduce.OutputFormat<K,V> mapreduceOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
AbstractID intermediateDataSetId
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeSerializer<T> serializer
TypeSerializer<T> typeSerializer
int id
SplittableIterator<T> source
PrintSinkOutputWriter<IN> writer
int arity
TypeInformation<T>[] fieldTypeInfos
int[] fieldPosMap
boolean emptyColumnAsNull
String partitionMarker
String charsetName
String charsetName
String charsetName
boolean skipInvalidLines
TupleSerializerBase<T> tupleSerializer
TypeSerializer<T> serializer
TypeSerializer<T> serializer
int[] fields
boolean[] isFromFirst
Tuple outTuple
CombineFunction<IN,OUT> wrappedFunction
RichGroupReduceFunction<IN,OUT> wrappedFunction
KeySelector<IN,KEY> keySelector1
KeySelector<IN,KEY> keySelector2
Tuple3<T0,T1,T2> tuple
Function wrappedFunction
long trueCount
long falseCount
long nullCount
double value
double delta
double max
double min
CompensatedSum sum
float max
float min
CompensatedSum sum
int max
int min
int sum
long max
long min
long nonMissingCount
long nullCount
long nanCount
long infinityCount
Aggregator<T,R> min
Aggregator<T,R> max
Aggregator<T,R> sum
CompensatedSum mean
CompensatedSum m2
The algorithm is described in: "Scalable and Numerically Stable Descriptive Statistics in SystemML", Tian et al, International Conference on Data Engineering 2012
long nonNullCount
long nullCount
short max
short min
short sum
long nonNullCount
long nullCount
long emptyCount
int minStringLength
int maxStringLength
CompensatedSum meanLength
Aggregator<T,R>[] columnAggregators
Aggregator<T,R> aggregator
private Object readResolve() throws ObjectStreamException
ObjectStreamException
Object f0
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f24
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
TypeInformation<T> leftType
TypeInformation<T> rightType
TypeInformation<T> elementTypeInfo
TypeInformation<T> keyTypeInfo
TypeInformation<T> valueTypeInfo
String functionName
InvalidTypesException typeException
Class<T> arrayType
TypeInformation<T> componentInfo
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeInformation<T> type
PojoField[] fields
int totalFields
String[] fieldNames
boolean[] comparatorOrders
String[] fieldNames
TypeInformation<T>[] types
int totalFields
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
boolean ascending
Class<T> type
TypeSerializer<T> serializer
TypeComparator<T>[] comparators
Class<T> registeredClass
This can be a dummy class KryoRegistrationSerializerConfigSnapshot.DummyRegisteredClass
if the class no longer exists
when this registration instance was restored.
Class<T> serializerClass
KryoRegistration.SerializerDefinitionType.CLASS
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if the serializer class no
longer exists when this registration instance was restored.
ExecutionConfig.SerializableSerializer<T extends com.esotericsoftware.kryo.Serializer<?> & Serializable> serializableSerializerInstance
KryoRegistration.SerializerDefinitionType.INSTANCE
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if the serializer class no
longer exists or is no longer valid when this registration instance was restored.
KryoRegistration.SerializerDefinitionType serializerDefinitionType
TypeSerializer<T> originalSerializer
byte[] padding
TypeComparator<T> wrappedComparator
boolean order
int flatFields
boolean nullReference
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T> serializer
Class<T> type
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> clazz
TypeSerializer<T>[] fieldSerializers
int numFields
LinkedHashMap<K,V> registeredClasses
TypeSerializer<T>[] registeredSerializers
ExecutionConfig executionConfig
int arity
int[] keyPositions
NullAwareComparator<T>[] comparators
TypeSerializer<T>[] serializers
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
boolean[] mask1
boolean[] mask2
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
boolean legacyModeEnabled
int legacyOffset
TypeSerializer<T>[] fieldSerializers
int arity
LinkedHashMap<K,V> positionByName
TypeComparator<T> comparator
TypeSerializer<T> serializer
boolean firstSerializer
Class<T> clazz
int[] keyPositions
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T>[] serializers
Class<T> tupleClass
TypeSerializer<T>[] fieldSerializers
int arity
int length
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Currently, we only have one single registration for the value type. Nevertheless, we keep this information here for future compatibility.
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
LinkedHashMap<K,V> defaultSerializers
LinkedHashMap<K,V> defaultSerializerClasses
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Class<T> type
LinkedHashMap<K,V> registeredTypesWithSerializers
LinkedHashMap<K,V> registeredTypesWithSerializerClasses
LinkedHashSet<E> registeredTypes
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Object[] extractedKeys
Class<T> clazz
TypeInformation<T>[] typeParamTypeInfos
scala.collection.Seq<A> org$apache$flink$api$scala$typeutils$CaseClassTypeInfo$$fieldTypes
scala.collection.Seq<A> fieldNames
String REGEX_INT_FIELD
String REGEX_STR_FIELD
String REGEX_FIELD
String REGEX_NESTED_FIELDS
String REGEX_NESTED_FIELDS_WILDCARD
Pattern PATTERN_NESTED_FIELDS
Pattern PATTERN_NESTED_FIELDS_WILDCARD
Pattern PATTERN_INT_FIELD
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
Class<T> clazz
TypeInformation<T> leftTypeInfo
TypeInformation<T> rightTypeInfo
boolean ascComp
IntComparator intComparator
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
scala.Enumeration enum
private Object readResolve()
TypeSerializer<T> elemSerializer
private Object readResolve()
boolean ascending
TypeComparator<T> typeComparator
scala.Option<A> reference
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
private Object readResolve()
byte ZeroInByte
byte OneInByte
TypeInformation<T> elemTypeInfo
private void readObject(ObjectInputStream arg0)
private Object readResolve()
private void readObject(ObjectInputStream arg0)
TypeSerializer<T> elementSerializer
String cbfCode
int classLoaderHash
scala.ref.WeakReference<T> classLoaderRef
String cbfCode
private Object readResolve()
ClassLoader classLoader
String code
private Object readResolve()
private Object readResolve()
org.apache.flink.shaded.guava18.com.google.common.cache.Cache<K,V> CACHE
Class<T> clazz
TypeInformation<T> elementTypeInfo
TypeSerializer<T> elemSerializer
TypeSerializer<T> throwableSerializer
private Object readResolve()
TypeInformation<T> elemTypeInfo
private Object readResolve()
org.slf4j.Logger logger
String query
ClusterBuilder builder
String insertQuery
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> inputClass
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> outputClass
String latchFilePath
long numRecords
long numKeys
long recordsPerPartition
long keysPerPartition
long recordCnt
int partitionId
boolean infinite
PatternFlatSelectFunction<IN,OUT> flatSelectFunction
PatternSelectFunction<IN,OUT> selectFunction
PatternFlatTimeoutFunction<IN,OUT> flatTimeoutFunction
OutputTag<T> timedOutPartialMatchesTag
PatternTimeoutFunction<IN,OUT> timeoutFunction
OutputTag<T> timedOutPartialMatchesTag
int[] deweyNumber
TypeSerializer<T> sharedBufferSerializer
TypeSerializer<T> eventSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> versionSerializer
NFAStateSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> nodeIdSerializer
TypeSerializer<T> eventIdSerializer
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeSerializer<T> versionSerializer
String name
State.StateType stateType
Collection<E> stateTransitions
StateTransitionAction action
State<T> sourceState
State<T> targetState
IterativeCondition<T> condition
TypeSerializer<T> elementSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> eventIdSerializer
NodeId.NodeIdSerializer.readObject(ObjectInputStream)
.private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> nodeIdSerializer
SharedBufferEdge.SharedBufferEdgeSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> deweyNumberSerializer
ListSerializer<T> edgesSerializer
TypeSerializer<T> edgeSerializer
boolean isProcessingTime
TypeSerializer<T> inputSerializer
NFACompiler.NFAFactory<T> nfaFactory
EventComparator<T> comparator
OutputTag<T> lateDataOutputTag
OutputTag
to use for late arriving events. Elements with timestamp smaller than the
current watermark will be emitted to this.AfterMatchSkipStrategy afterMatchSkipStrategy
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T> original
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T>[] nestedConditions
ApplicationStatus status
Configuration backingConfig
String prefix
long bytes
int sourceIndex
int sourceIndex
Source<T,SplitT extends SourceSplit,EnumChkT> source
boolean finalSource
org.apache.flink.connector.file.sink.FileSink.BucketsBuilder<IN,T extends org.apache.flink.connector.file.sink.FileSink.BucketsBuilder<IN,T>> bucketsBuilder
Path basePath
long bucketCheckInterval
BulkWriter.Factory<T> writerFactory
FileWriterBucketFactory<IN> bucketFactory
BucketAssigner<IN,BucketID> bucketAssigner
CheckpointRollingPolicy<IN,BucketID> rollingPolicy
OutputFileConfig outputFileConfig
Path basePath
long bucketCheckInterval
Encoder<IN> encoder
FileWriterBucketFactory<IN> bucketFactory
BucketAssigner<IN,BucketID> bucketAssigner
RollingPolicy<IN,BucketID> rollingPolicy
OutputFileConfig outputFileConfig
InProgressFileWriter.PendingFileRecoverable pendingFile
InProgressFileWriter.InProgressFileRecoverable inProgressFileToCleanup
Path[] inputPaths
FileEnumerator.Provider enumeratorFactory
FileSplitAssigner.Provider assignerFactory
BulkFormat<T,SplitT extends FileSourceSplit> readerFormat
ContinuousEnumerationSettings continuousEnumerationSettings
java.time.Duration discoveryInterval
String id
Path filePath
long offset
long length
String[] hostnames
CheckpointedPosition readerPosition
FileRecordFormat<T> fileFormat
StreamFormat<T> streamFormat
String charsetName
long offset
long recordsAfterOffset
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
boolean lookupAsync
long bufferFlushMaxSizeInBytes
long bufferFlushMaxRows
long bufferFlushIntervalMillis
Integer parallelism
String hTableName
byte[] serializedConfig
long bufferFlushMaxSizeInBytes
long bufferFlushMaxMutations
long bufferFlushIntervalMillis
HBaseMutationConverter<T> mutationConverter
AtomicReference<V> failureThrowable
BufferedMutator.ExceptionListener
if a Throwable
was thrown.
Errors will be checked and rethrown before processing each input element, and when the sink is closed.
HBaseTableSchema schema
HBaseTableSchema schema
String nullStringLiteral
String hTableName
byte[] serializedConfig
HBaseTableSchema hbaseTableSchema
String hTableName
byte[] serializedConfig
HBaseTableSchema hbaseTableSchema
String nullStringLiteral
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
byte[] tableName
byte[] startRow
byte[] endRow
boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] currentRow
long scannedRows
byte[] serializedConfig
String tableName
HBaseTableSchema schema
String nullStringLiteral
String tableName
HBaseTableSchema schema
boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] currentRow
long scannedRows
byte[] serializedConfig
String hTableName
byte[] serializedConfig
HBaseTableSchema hbaseTableSchema
String nullStringLiteral
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
String tableName
HBaseTableSchema schema
String nullStringLiteral
String tableName
HBaseTableSchema schema
boolean discoverAndRollbackOnRecovery
int maxCommitAttempts
boolean allowOutOfOrderCommits
Integer timeoutSec
boolean transactionPerConnection
long batchIntervalMs
int batchSize
int maxRetries
JdbcConnectionProvider connectionProvider
String queryTemplate
int resultSetType
int resultSetConcurrency
RowTypeInfo rowTypeInfo
int fetchSize
Boolean autoCommit
boolean hasNext
Object[][] parameterValues
JdbcConnectionProvider connectionProvider
AbstractJdbcOutputFormat<T> outputFormat
JdbcExecutionOptions executionOptions
JdbcBatchingOutputFormat.StatementExecutorFactory<T extends JdbcBatchStatementExecutor<?>> statementExecutorFactory
JdbcBatchingOutputFormat.RecordExtractor<F,T> jdbcRecordExtractor
JdbcConnectionOptions jdbcOptions
RowType rowType
org.apache.flink.connector.jdbc.internal.converter.AbstractJdbcRowConverter.JdbcDeserializationConverter[] toInternalConverters
org.apache.flink.connector.jdbc.internal.converter.AbstractJdbcRowConverter.JdbcSerializationConverter[] toExternalConverters
LogicalType[] fieldTypes
String[] fieldNames
String[] keyFields
String tableName
JdbcDialect dialect
String query
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
String tableName
JdbcDialect dialect
Integer parallelism
JdbcOptions jdbcOptions
JdbcExecutionOptions executionOptions
JdbcDmlOptions dmlOptions
TypeInformation<T> rowDataTypeInformation
DataType[] fieldDataTypes
String query
JdbcConnectionProvider connectionProvider
TypeInformation<T>[] keyTypes
int[] keySqlTypes
String[] fieldNames
String[] keyNames
TypeInformation<T>[] fieldTypes
int[] outputSqlTypes
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
JdbcConnectionProvider connectionProvider
int fetchSize
Boolean autoCommit
Object[][] parameterValues
String queryTemplate
int resultSetType
int resultSetConcurrency
JdbcRowConverter rowConverter
TypeInformation<T> rowDataTypeInfo
String query
JdbcConnectionProvider connectionProvider
DataType[] keyTypes
String[] keyNames
long cacheMaxSize
long cacheExpireMs
int maxRetryTimes
JdbcDialect jdbcDialect
JdbcRowConverter jdbcRowConverter
JdbcRowConverter lookupKeyRowConverter
TypeSerializer<T> xidSerializer
XaFacade xaFacade
org.apache.flink.connector.jdbc.xa.XaGroupOps xaGroupOps
XidGenerator xidGenerator
JdbcBatchingOutputFormat<In,JdbcIn,JdbcExec extends JdbcBatchStatementExecutor<JdbcIn>> outputFormat
org.apache.flink.connector.jdbc.xa.XaSinkStateHandler stateHandler
JdbcExactlyOnceOptions options
Xid xid
TypeSerializer<T> xidSerializer
TypeSerializer<T> checkpointAndXidSerializer
KafkaSubscriber subscriber
OffsetsInitializer startingOffsetsInitializer
OffsetsInitializer stoppingOffsetsInitializer
Boundedness boundedness
KafkaRecordDeserializationSchema<T> deserializationSchema
Properties props
JobConfWrapper jobConfWrapper
DataFormatConverters.DataFormatConverter<Internal,External>[] partitionConverters
HiveObjectConversion[] hiveObjectConversions
HiveObjectConversion[] partColConversions
JobConfWrapper jobConfWrapper
List<E> partitionKeys
ContinuousPartitionFetcher<P,T extends Comparable<T>> fetcher
HiveTableSource.HiveContinuousPartitionFetcherContext<T extends Comparable<T>> fetcherContext
ObjectPath tablePath
List<E> partitions
JobConfWrapper jobConfWrapper
JobConfWrapper conf
String hiveVersion
String database
String tableName
CachedSerializedValue<T> storageDescriptor
Map<K,V> partitionSpec
Properties tableProps
TypeSerializer<T> typeSerializer
Comparable<T> consumeStartOffset
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
JobConfWrapper jobConfWrapper
List<E> partitionKeys
String[] fieldNames
DataType[] fieldTypes
String hiveVersion
HiveShim hiveShim
RowType producedRowType
boolean useMapRedReader
PartitionFieldExtractor<T extends FileSourceSplit> partitionFieldExtractor
CachedSerializedValue<T> sd
Properties properties
JobConfWrapper jobConfWrapper
List<E> partitionKeys
String[] fieldNames
DataType[] fieldTypes
String hiveVersion
HiveShim shim
RowType producedRowType
boolean useMapRedReader
JobConfWrapper jobConfWrapper
String hiveVersion
ObjectPath tablePath
DataType[] fieldTypes
String[] fieldNames
List<E> partitionKeys
int[] selectedFields
boolean useMapRedReader
ObjectPath tablePath
HiveShim hiveShim
JobConfWrapper confWrapper
List<E> partitionKeys
DataType[] fieldTypes
String[] fieldNames
Configuration configuration
String defaultPartitionName
ConsumeOrder consumeOrder
HiveTablePartition hiveTablePartition
HiveTableInputFormat inputFormat
HiveTablePartition hiveTablePartition
HiveTablePartition hiveTablePartition
HiveWriterFactory factory
HiveWriterFactory factory
Class<T> hiveOutputFormatClz
CachedSerializedValue<T> serDeInfo
String[] allColumns
DataType[] allTypes
String[] partitionColumns
Properties tableProperties
JobConfWrapper confWrapper
HiveShim hiveShim
boolean isCompressed
File[] localRocksDbDirectories
PredefinedOptions predefinedOptions
RocksDBOptionsFactory rocksDbOptionsFactory
TernaryBoolean enableIncrementalCheckpointing
int numberOfTransferThreads
RocksDBMemoryConfiguration memoryConfiguration
EmbeddedRocksDBStateBackend.PriorityQueueStateType priorityQueueStateType
RocksDBNativeMetricOptions defaultMetricOptions
long writeBatchSize
RocksDBWriteBatchWrapper
, default value
2mb.Boolean useManagedMemory
MemorySize fixedMemoryPerSlot
Double writeBufferRatio
Double highPriorityPoolRatio
Boolean usePartitionedIndexFilters
EmbeddedRocksDBStateBackend rocksDBStateBackend
StateBackend checkpointStreamBackend
Path file
long start
long length
URI uri
int partitionNumber
int totalNumberOfPartitions
int splitNumber
String[] hostnames
SerializableSupplier<T> serializerSupplier
byte[] buffer
int end
int position
int id
double x
double y
Collection<E> centroids
List<E> tasks
FileCopyTaskInputSplit curInputSplit
FileCopyTask task
int splitNumber
EnumTrianglesDataTypes.Edge outEdge
double dampening
double randomJump
EmptyFieldsCountAccumulator.VectorAccumulator emptyFieldCounter
int dataInfoCase_
Object dataInfo_
int dataType_
int outputMode_
byte memoizedIsInitialized
int windowType_
boolean isTimeWindow_
long windowSlide_
long windowSize_
long windowGap_
boolean isRowTime_
int timeFieldIndex_
long allowedLateness_
List<E> namedProperties_
int namedPropertiesMemoizedSerializedSize
Object shiftTimezone_
byte memoizedIsInitialized
int inputCase_
Object input_
byte memoizedIsInitialized
int windowType_
long lowerBoundary_
long upperBoundary_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int precision_
int scale_
byte memoizedIsInitialized
Object name_
Object description_
FlinkFnApi.Schema.FieldType type_
byte memoizedIsInitialized
int typeInfoCase_
Object typeInfo_
int typeName_
boolean nullable_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType keyType_
FlinkFnApi.Schema.FieldType valueType_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int length_
byte memoizedIsInitialized
int precision_
byte memoizedIsInitialized
int typeInfoCase_
Object typeInfo_
int typeName_
byte memoizedIsInitialized
FlinkFnApi.TypeInfo keyType_
FlinkFnApi.TypeInfo valueType_
byte memoizedIsInitialized
Object fieldName_
FlinkFnApi.TypeInfo fieldType_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType elementType_
byte memoizedIsInitialized
FlinkFnApi.Schema.FieldType keyType_
FlinkFnApi.Schema.FieldType valueType_
byte memoizedIsInitialized
List<E> udfs_
boolean metricEnabled_
com.google.protobuf.Internal.IntList grouping_
int groupingMemoizedSerializedSize
boolean generateUpdateBefore_
FlinkFnApi.Schema.FieldType keyType_
int indexOfCountStar_
boolean stateCleaningEnabled_
int stateCacheSize_
int mapStateReadCacheSize_
int mapStateWriteCacheSize_
boolean countStarInserted_
FlinkFnApi.GroupWindow groupWindow_
byte memoizedIsInitialized
int functionType_
FlinkFnApi.UserDefinedDataStreamFunction.RuntimeContext runtimeContext_
com.google.protobuf.ByteString payload_
boolean metricEnabled_
FlinkFnApi.TypeInfo keyTypeInfo_
byte memoizedIsInitialized
Object taskName_
Object taskNameWithSubtasks_
int numberOfParallelSubtasks_
int maxNumberOfParallelSubtasks_
int indexOfThisSubtask_
int attemptNumber_
List<E> jobParameters_
boolean inBatchExecutionMode_
byte memoizedIsInitialized
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
DeserializationSchema<T> nestedSchema
GenericRecord
. *TypeInformation<T> typeInfo
AvroToRowDataConverters.AvroToRowDataConverter runtimeConverter
SerializationSchema<T> nestedSchema
GenericRecord
into bytes. *RowType rowType
RowDataToAvroConverters.RowDataToAvroConverter runtimeConverter
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
AvroBuilder<T> avroBuilder
DataFileWriter
.SchemaCoder.SchemaCoderProvider schemaCoderProvider
SchemaCoder.SchemaCoderProvider schemaCoderProvider
SchemaCoder schemaCoder
AvroRowDataDeserializationSchema avroDeserializer
TypeInformation<T> producedTypeInfo
RowData
.AvroRowDataSerializationSchema avroSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream ois) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
RowType formatRowType
DataType[] fieldTypes
String[] fieldNames
int[] selectFields
List<E> partitionKeys
String defaultPartValue
long limit
int[] csvSelectFieldToProjectFieldMapping
int[] csvSelectFieldToCsvFieldMapping
boolean ignoreParseErrors
TypeInformation<T> resultTypeInfo
CsvToRowDataConverters.CsvToRowDataConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader objectReader
CsvSchema
.boolean ignoreParseErrors
RowType rowType
RowDataToCsvConverters.RowDataToCsvConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvMapper csvMapper
JsonNode
into bytes.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectWriter objectWriter
CsvSchema
.TypeInformation<T> typeInfo
org.apache.flink.formats.csv.CsvRowDeserializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader objectReader
CsvSchema
.boolean ignoreParseErrors
RowTypeInfo typeInfo
org.apache.flink.formats.csv.CsvRowSerializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvMapper csvMapper
JsonNode
into bytes.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectWriter objectWriter
CsvSchema
.boolean ignoreParseErrors
TypeInformation<T>[] fieldTypes
String[] selectedFieldNames
boolean ignoreParseErrors
Path[] filePaths
TypeInformation<T>[] fieldTypes
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
boolean ignoreParseErrors
int[] selectedFields
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
boolean failOnMissingField
boolean ignoreParseErrors
TypeInformation<T> resultTypeInfo
RowData
.JsonToRowDataConverters.JsonToRowDataConverter runtimeConverter
JsonNode
s into objects of Flink SQL internal data
structures.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
TimestampFormat timestampFormat
RowType rowType
RowDataToJsonConverters.RowDataToJsonConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
TimestampFormat timestampFormat
JsonOptions.MapNullKeyMode mapNullKeyMode
String mapNullKeyLiteral
boolean encodeDecimalAsPlainNumber
RowTypeInfo typeInfo
boolean failOnMissingField
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
org.apache.flink.formats.json.JsonRowDeserializationSchema.DeserializationRuntimeConverter runtimeConverter
boolean ignoreParseErrors
RowTypeInfo typeInfo
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
org.apache.flink.formats.json.JsonRowSerializationSchema.SerializationRuntimeConverter runtimeConverter
boolean failOnMissingField
boolean ignoreParseErrors
TimestampFormat timestampFormat
TimestampFormat timestampFormat
JsonOptions.MapNullKeyMode mapNullKeyMode
String mapNullKeyLiteral
JsonRowDataDeserializationSchema jsonDeserializer
boolean hasMetadata
org.apache.flink.formats.json.canal.CanalJsonDeserializationSchema.MetadataConverter[] metadataConverters
TypeInformation<T> producedTypeInfo
TypeInformation
of the produced RowData
(physical + meta data).String database
String table
boolean ignoreParseErrors
List<E> fieldNames
int fieldCount
Pattern databasePattern
Pattern tablePattern
JsonRowDataSerializationSchema jsonSerializer
JsonRowDataDeserializationSchema jsonDeserializer
boolean hasMetadata
org.apache.flink.formats.json.debezium.DebeziumJsonDeserializationSchema.MetadataConverter[] metadataConverters
TypeInformation<T> producedTypeInfo
TypeInformation
of the produced RowData
(physical + meta data).boolean schemaInclude
boolean ignoreParseErrors
JsonRowDataSerializationSchema jsonSerializer
JsonRowDataDeserializationSchema jsonDeserializer
TypeInformation<T> resultTypeInfo
RowData
. *boolean ignoreParseErrors
int fieldCount
JsonRowDataSerializationSchema jsonSerializer
TimestampFormat timestampFormat
String avroSchemaString
RowType producedType
boolean skipWrongSchemaFileSplit
boolean skipCorruptedRecord
boolean skipThisSplit
TypeInformation<T>[] fieldTypes
String[] fieldNames
org.apache.parquet.filter2.predicate.FilterPredicate filterPredicate
Class<T> pojoTypeClass
TypeSerializer<T> typeSerializer
SerializableConfiguration hadoopConfig
String[] projectedFields
LogicalType[] projectedTypes
ColumnBatchFactory<SplitT extends FileSourceSplit> batchFactory
int batchSize
boolean isUtcTimestamp
boolean isCaseSensitive
ParquetBuilder<T> writerBuilder
RowType rowType
SerializableConfiguration configuration
boolean utcTimestamp
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration conf
org.apache.flink.formats.sequencefile.SerializableHadoopConfiguration serializableHadoopConfig
Class<T> keyClass
Class<T> valueClass
String compressionCodecName
org.apache.hadoop.io.SequenceFile.CompressionType compressionType
String id
long count
long checksum
MurmurHash hasher
Object vertexId0
long offset
double x
double y
double beta
long srcId
long offset
long length
long numberOfVertices
IterationRuntimeContext runtimeContext
long numberOfVertices
IterationRuntimeContext runtimeContext
TypeInformation<T> typeInformation
TypeInformation<T> typeInformation
Either.Right<L,R> nullValue
ByteValue bitmask
DoubleValue hubScore
DoubleValue authorityScore
DoubleValue pageRankScore
FloatValue adamicAdarScore
Vertex<K,V> outVertex
Tuple2<T0,T1> outMsg
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
org.apache.flink.graph.pregel.ComputeFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
boolean setNewVertexValueCalled
Object first
long numberOfVertices
boolean optDegrees
IterationRuntimeContext runtimeContext
Collector<T> out
Collector<T> outWithDegrees
Vertex<K,V> outVal
Vertex<K,V> outValWithDegrees
long inDegree
long outDegree
boolean setNewVertexValueCalled
long numberOfVertices
EdgeDirection direction
Tuple2<T0,T1> outValue
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
Object vertexId
org.apache.flink.graph.spargel.ScatterFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
long inDegree
long outDegree
boolean isBounded
byte[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.ByteValueArray.ReadIterator iterator
boolean ascendingComparison
ByteValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
char[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.CharValueArray.ReadIterator iterator
boolean ascendingComparison
CharValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
double[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.DoubleValueArray.ReadIterator iterator
boolean ascendingComparison
DoubleValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
float[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.FloatValueArray.ReadIterator iterator