int numKeys
int idlenessMs
int recordsToEmit
int numRecordsEmitted
boolean canceled
public void readExternal(ObjectInput in) throws IOException
IOException
public void writeExternal(ObjectOutput out) throws IOException
IOException
byte[] bytes
boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] currentRow
long scannedRows
String tableName
HBaseTableSchema schema
byte[][] families
byte[][][] qualifiers
int[][] types
Row resultRow
Row[] familyRows
byte[] tableName
byte[] startRow
byte[] endRow
ExecutionMode executionMode
ExecutionConfig.ClosureCleanerLevel closureCleanerLevel
int parallelism
int maxParallelism
int numberOfExecutionRetries
boolean forceKryo
boolean disableGenericTypes
boolean objectReuse
boolean autoTypeRegistrationEnabled
boolean forceAvro
CodeAnalysisMode codeAnalysisMode
boolean printProgressDuringExecution
long autoWatermarkInterval
long latencyTrackingInterval
boolean isLatencyTrackingConfigured
long executionRetryDelay
RestartStrategies.RestartStrategyConfiguration restartStrategyConfiguration
long taskCancellationIntervalMillis
long taskCancellationTimeoutMillis
boolean useSnapshotCompression
boolean failTaskOnCheckpointError
InputDependencyConstraint defaultInputDependencyConstraint
ExecutionConfig.GlobalJobParameters globalJobParameters
LinkedHashMap<K,V> registeredTypesWithKryoSerializers
LinkedHashMap<K,V> registeredTypesWithKryoSerializerClasses
LinkedHashMap<K,V> defaultKryoSerializers
LinkedHashMap<K,V> defaultKryoSerializerClasses
LinkedHashSet<E> registeredKryoTypes
LinkedHashSet<E> registeredPojoTypes
com.esotericsoftware.kryo.Serializer<T> serializer
long count
double sum
double localValue
double max
double min
int localValue
int max
int min
long localValue
long max
long min
DoubleValue wrapper
double sum
long sum
TypeComparator<T> typeComparator
Object[][] boundaries
TypeComparator<T>[] flatComparators
Object[] keys
boolean target
String sinkIdentifier
long blockSize
long readRecords
long blockSize
Path filePath
FileInputFormat.supportsMultiPaths()
and
use FileInputFormat.getFilePaths()
and FileInputFormat.setFilePaths(Path...)
.Path[] filePaths
long minSplitSize
int numSplits
long openTimeout
boolean unsplittable
boolean enumerateNestedFiles
FilePathFilter filesFilter
Path outputFilePath
FileSystem.WriteMode writeMode
FileOutputFormat.OutputDirectoryMode outputDirectoryMode
boolean lineDelimiterIsLinebreak
Class<T>[] fieldTypes
boolean[] fieldIncluded
byte[] fieldDelim
String fieldDelimString
boolean lenient
boolean skipFirstLineAsHeader
boolean quotedStringParsing
byte quoteCharacter
byte[] commentPrefix
String commentPrefixString
int partitionNumber
InputFormat<OT,T extends InputSplit> replicatedIF
long globalRateBytesPerSecond
long localRateBytesPerSecond
RuntimeContext runtimeContext
org.apache.flink.shaded.guava18.com.google.common.util.concurrent.RateLimiter rateLimiter
Map<K,V> fieldMapping1
Map<K,V> fieldMapping2
FieldSet readFields1
FieldSet readFields2
double cpuCores
int heapMemoryInMB
int directMemoryInMB
int nativeMemoryInMB
int stateSizeInMB
Map<K,V> extendedResources
long count
BulkIterationBase.TerminationCriterionAggregator aggregator
Object userCodeObject
String name
double value
Resource.ResourceAggregateType resourceAggregateType
int restartAttempts
Time delayBetweenAttemptsInterval
TypeInformation<T> type
DeserializationSchema
.String charsetName
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeInformation<T> typeInfo
TypeInformationSerializationSchema.getProducedType()
.TypeSerializer<T> serializer
AggregateFunction<IN,ACC,OUT> aggFunction
FoldFunction<O,T> foldFunction
ReduceFunction<T> reduceFunction
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
String name
AtomicReference<V> serializerAtomicReference
StateDescriptor.initializeSerializerUnlessSet(ExecutionConfig)
method
is called.TypeInformation<T> typeInfo
String queryableStateName
StateTtlConfig ttlConfig
StateTtlConfig.UpdateType updateType
StateTtlConfig.StateVisibility stateVisibility
StateTtlConfig.TimeCharacteristic timeCharacteristic
Time ttl
StateTtlConfig.CleanupStrategies cleanupStrategies
int cleanupSize
boolean runCleanupForEveryRecord
long queryTimeAfterNumEntries
TimeUnit unit
long size
Class<T> arrayClass
TypeInformation<T> componentInfo
TypeSerializer<T>[] fieldSerializers
CompositeSerializer.PrecomputedParameters precomputed
boolean immutableTargetType
boolean immutable
int length
boolean stateful
TypeComparator<T> comparator1
TypeComparator<T> comparator2
TypeComparator<T>[] comparators1
TypeComparator<T>[] comparators2
Object[] referenceKeyFields
Object[] candidateKeyFields
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
org.apache.flink.api.common.typeutils.SingleThreadAccessCheckingTypeSerializer.SingleThreadAccessChecker singleThreadAccessChecker
TypeSerializer<T> originalSerializer
byte[] actualBytes
Throwable originalError
boolean ascendingComparison
TypeComparator<T>[] comparators
boolean ascendingComparison
BooleanValue reference
BooleanValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
ByteValue reference
ByteValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
CharValue reference
CharValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
DoubleValue reference
DoubleValue tempReference
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> enumClass
Map<K,V> valueToOrdinal
Enum.ordinal()
.
This allows us to maintain backwards compatibility for previous serialized data in the case that the
order of enum constants was changed or new constants were added.
On a fresh start with no reconfiguration, the ordinals would simply be identical to the enum constants actual ordinals. Ordinals may change after reconfiguration.
Enum<E extends Enum<E>>[] values
EnumSerializer.valueToOrdinal
map.
Serves as a bidirectional map to have fast access from ordinal to value. May be reordered after reconfiguration.boolean ascendingComparison
FloatValue reference
FloatValue tempReference
TypeComparator<T>[] comparators
Class<T> componentClass
TypeSerializer<T> componentSerializer
boolean ascendingComparison
IntValue reference
IntValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> elementSerializer
boolean ascendingComparison
LongValue reference
LongValue tempReference
TypeComparator<T>[] comparators
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeComparator<T>[] comparators
boolean ascendingComparison
ShortValue reference
ShortValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
StringValue reference
StringValue tempReference
TypeComparator<T>[] comparators
TypeComparator<T>[] comparators
boolean ascending
BasicTypeComparator<T extends Comparable<T>> comparator
long count
long checksum
String id
long counter
long checksum
String id
TypeSerializer<T> serializer
SerializedListAccumulator<T> accumulator
String id
long counter
int count
TextOutputFormat.TextFormatter<IN> formatter
boolean withReplacement
int numSample
long seed
boolean withReplacement
int numSample
long seed
boolean withReplacement
double fraction
long seed
int[] fields
int[] fields
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapred.JobConf jobConf
org.apache.hadoop.mapred.OutputFormat<K,V> mapredOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapreduce.InputFormat<K,V> mapreduceInputFormat
Class<T> keyClass
Class<T> valueClass
Configuration configuration
boolean fetched
boolean hasNext
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration configuration
org.apache.hadoop.mapreduce.OutputFormat<K,V> mapreduceOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeSerializer<T> serializer
TypeSerializer<T> typeSerializer
int id
SplittableIterator<T> source
PrintSinkOutputWriter<IN> writer
int arity
TypeInformation<T>[] fieldTypeInfos
int[] fieldPosMap
boolean emptyColumnAsNull
String partitionMarker
String charsetName
String charsetName
String charsetName
boolean skipInvalidLines
TupleSerializerBase<T> tupleSerializer
TypeSerializer<T> serializer
TypeSerializer<T> serializer
String username
String password
String drivername
String dbURL
String queryTemplate
int resultSetType
int resultSetConcurrency
RowTypeInfo rowTypeInfo
int fetchSize
boolean hasNext
Object[][] parameterValues
String username
String password
String drivername
String dbURL
String query
int batchInterval
Connection dbConn
PreparedStatement upload
int batchCount
int[] typesArray
int[] fields
boolean[] isFromFirst
Tuple outTuple
CombineFunction<IN,OUT> wrappedFunction
RichGroupReduceFunction<IN,OUT> wrappedFunction
KeySelector<IN,KEY> keySelector1
KeySelector<IN,KEY> keySelector2
Tuple3<T0,T1,T2> tuple
Function wrappedFunction
long trueCount
long falseCount
long nullCount
double value
double delta
double max
double min
CompensatedSum sum
float max
float min
CompensatedSum sum
int max
int min
int sum
long max
long min
long nonMissingCount
long nullCount
long nanCount
long infinityCount
Aggregator<T,R> min
Aggregator<T,R> max
Aggregator<T,R> sum
CompensatedSum mean
CompensatedSum m2
The algorithm is described in: "Scalable and Numerically Stable Descriptive Statistics in SystemML", Tian et al, International Conference on Data Engineering 2012
long nonNullCount
long nullCount
short max
short min
short sum
long nonNullCount
long nullCount
long emptyCount
int minStringLength
int maxStringLength
CompensatedSum meanLength
Aggregator<T,R>[] columnAggregators
Aggregator<T,R> aggregator
private Object readResolve() throws ObjectStreamException
ObjectStreamException
Object f0
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f24
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
TypeInformation<T> leftType
TypeInformation<T> rightType
TypeInformation<T> elementTypeInfo
TypeInformation<T> keyTypeInfo
TypeInformation<T> valueTypeInfo
String functionName
InvalidTypesException typeException
Class<T> arrayType
TypeInformation<T> componentInfo
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeInformation<T> type
PojoField[] fields
int totalFields
String[] fieldNames
boolean[] comparatorOrders
String[] fieldNames
TypeInformation<T>[] types
int totalFields
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
boolean ascending
Class<T> type
TypeSerializer<T> serializer
TypeComparator<T>[] comparators
Class<T> registeredClass
This can be a dummy class KryoRegistrationSerializerConfigSnapshot.DummyRegisteredClass
if
the class no longer exists when this registration instance was restored.
Class<T> serializerClass
KryoRegistration.SerializerDefinitionType.CLASS
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if
the serializer class no longer exists when this registration instance was restored.
ExecutionConfig.SerializableSerializer<T extends com.esotericsoftware.kryo.Serializer<?> & Serializable> serializableSerializerInstance
KryoRegistration.SerializerDefinitionType.INSTANCE
.
This can be a dummy serializer KryoRegistrationSerializerConfigSnapshot.DummyKryoSerializerClass
if
the serializer class no longer exists or is no longer valid when this registration instance was restored.
KryoRegistration.SerializerDefinitionType serializerDefinitionType
TypeSerializer<T> originalSerializer
byte[] padding
TypeComparator<T> wrappedComparator
boolean order
int flatFields
boolean nullReference
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T> serializer
Class<T> type
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> clazz
TypeSerializer<T>[] fieldSerializers
int numFields
LinkedHashMap<K,V> registeredClasses
TypeSerializer<T>[] registeredSerializers
ExecutionConfig executionConfig
int arity
int[] keyPositions
NullAwareComparator<T>[] comparators
TypeSerializer<T>[] serializers
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
boolean[] nullMask1
boolean[] nullMask2
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T>[] fieldSerializers
int arity
TypeComparator<T> comparator
TypeSerializer<T> serializer
boolean firstSerializer
Class<T> clazz
int[] keyPositions
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T>[] serializers
Class<T> tupleClass
TypeSerializer<T>[] fieldSerializers
int arity
int length
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Currently, we only have one single registration for the value type. Nevertheless, we keep this information here for future compatibility.
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
LinkedHashMap<K,V> defaultSerializers
LinkedHashMap<K,V> defaultSerializerClasses
LinkedHashMap<K,V> kryoRegistrations
This map serves as a preview of the final registration result of the Kryo instance, taking into account registration overwrites.
Class<T> type
LinkedHashMap<K,V> registeredTypesWithSerializers
LinkedHashMap<K,V> registeredTypesWithSerializerClasses
LinkedHashSet<E> registeredTypes
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Object[] extractedKeys
Class<T> clazz
TypeInformation<T>[] typeParamTypeInfos
scala.collection.Seq<A> org$apache$flink$api$scala$typeutils$CaseClassTypeInfo$$fieldTypes
scala.collection.Seq<A> fieldNames
String REGEX_INT_FIELD
String REGEX_STR_FIELD
String REGEX_FIELD
String REGEX_NESTED_FIELDS
String REGEX_NESTED_FIELDS_WILDCARD
Pattern PATTERN_NESTED_FIELDS
Pattern PATTERN_NESTED_FIELDS_WILDCARD
Pattern PATTERN_INT_FIELD
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
Class<T> clazz
TypeInformation<T> leftTypeInfo
TypeInformation<T> rightTypeInfo
boolean ascComp
IntComparator intComparator
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
scala.Enumeration enum
private Object readResolve()
TypeSerializer<T> elemSerializer
private Object readResolve()
boolean ascending
TypeComparator<T> typeComparator
scala.Option<A> reference
TypeComparator<T>[] getFlatComparators
boolean bitmap$0
private Object readResolve()
byte ZeroInByte
byte OneInByte
TypeInformation<T> elemTypeInfo
private void readObject(ObjectInputStream arg0)
private Object readResolve()
private void readObject(ObjectInputStream arg0)
TypeSerializer<T> elementSerializer
String cbfCode
int classLoaderHash
scala.ref.WeakReference<T> classLoaderRef
String cbfCode
private Object readResolve()
ClassLoader classLoader
String code
private Object readResolve()
private Object readResolve()
org.apache.flink.shaded.guava18.com.google.common.cache.Cache<K,V> CACHE
Class<T> clazz
TypeInformation<T> elementTypeInfo
TypeSerializer<T> elemSerializer
TypeSerializer<T> throwableSerializer
private Object readResolve()
TypeInformation<T> elemTypeInfo
private Object readResolve()
public void readExternal(ObjectInput in) throws IOException
IOException
public void writeExternal(ObjectOutput out) throws IOException
IOException
org.slf4j.Logger logger
String query
ClusterBuilder builder
String insertQuery
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> inputClass
ClusterBuilder builder
MapperOptions mapperOptions
Class<T> outputClass
long numRecords
long numKeys
long recordsPerPartition
long keysPerPartition
long recordCnt
int partitionId
boolean infinite
PatternFlatSelectFunction<IN,OUT> flatSelectFunction
PatternSelectFunction<IN,OUT> selectFunction
PatternFlatTimeoutFunction<IN,OUT> flatTimeoutFunction
OutputTag<T> timedOutPartialMatchesTag
PatternTimeoutFunction<IN,OUT> timeoutFunction
OutputTag<T> timedOutPartialMatchesTag
int[] deweyNumber
TypeSerializer<T> sharedBufferSerializer
TypeSerializer<T> eventSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> versionSerializer
NFAStateSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> nodeIdSerializer
TypeSerializer<T> eventIdSerializer
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
TypeSerializer<T> versionSerializer
String name
State.StateType stateType
Collection<E> stateTransitions
StateTransitionAction action
State<T> sourceState
State<T> targetState
IterativeCondition<T> condition
TypeSerializer<T> elementSerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> eventIdSerializer
NodeId.NodeIdSerializer.readObject(ObjectInputStream)
.private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> nodeIdSerializer
SharedBufferEdge.SharedBufferEdgeSerializer.readObject(ObjectInputStream)
.TypeSerializer<T> deweyNumberSerializer
ListSerializer<T> edgesSerializer
boolean isProcessingTime
TypeSerializer<T> inputSerializer
NFACompiler.NFAFactory<T> nfaFactory
long lastWatermark
EventComparator<T> comparator
OutputTag<T> lateDataOutputTag
OutputTag
to use for late arriving events. Elements with timestamp smaller than
the current watermark will be emitted to this.AfterMatchSkipStrategy afterMatchSkipStrategy
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T> original
IterativeCondition<T> left
IterativeCondition<T> right
IterativeCondition<T>[] nestedConditions
Configuration backingConfig
String prefix
long bytes
StateBackend checkpointStreamBackend
File[] localRocksDbDirectories
PredefinedOptions predefinedOptions
OptionsFactory optionsFactory
TernaryBoolean enableIncrementalCheckpointing
int numberOfTransferingThreads
TernaryBoolean enableTtlCompactionFilter
Note: User can still decide in state TTL configuration in state descriptor whether the filter is active for particular state or not.
RocksDBStateBackend.PriorityQueueStateType priorityQueueStateType
RocksDBNativeMetricOptions defaultMetricOptions
Path file
long start
long length
URI uri
int partitionNumber
int totalNumberOfPartitions
int splitNumber
String[] hostnames
byte[] buffer
int end
int position
int id
double x
double y
Collection<E> centroids
List<E> tasks
FileCopyTaskInputSplit curInputSplit
FileCopyTask task
int splitNumber
EnumTrianglesDataTypes.Edge outEdge
double dampening
double randomJump
double x
double y
double theta0
double theta1
Collection<E> parameters
LinearRegression.Params parameter
int count
EmptyFieldsCountAccumulator.VectorAccumulator emptyFieldCounter
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
private void readObject(ObjectInputStream inputStream) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream outputStream) throws IOException
IOException
SchemaCoder.SchemaCoderProvider schemaCoderProvider
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream ois) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
TypeInformation<T> typeInfo
org.apache.flink.formats.csv.CsvRowDeserializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader objectReader
CsvSchema
.boolean ignoreParseErrors
RowTypeInfo typeInfo
org.apache.flink.formats.csv.CsvRowSerializationSchema.RuntimeConverter runtimeConverter
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvMapper csvMapper
JsonNode
into bytes.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvSchema csvSchema
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectWriter objectWriter
CsvSchema
.org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
TypeInformation<T> typeInfo
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
boolean failOnMissingField
TypeInformation<T> typeInfo
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
SimpleDateFormat timeFormat
SimpleDateFormat timeFormatWithMillis
SimpleDateFormat timestampFormat
ParquetBuilder<T> writerBuilder
org.apache.flink.formats.sequencefile.SerializableHadoopConfiguration serializableHadoopConfig
Class<T> keyClass
Class<T> valueClass
String compressionCodecName
org.apache.hadoop.io.SequenceFile.CompressionType compressionType
String id
long count
long checksum
MurmurHash hasher
Object vertexId0
long offset
double x
double y
double beta
long srcId
long offset
long length
long numberOfVertices
IterationRuntimeContext runtimeContext
long numberOfVertices
IterationRuntimeContext runtimeContext
TypeInformation<T> typeInformation
TypeInformation<T> typeInformation
Either.Right<L,R> nullValue
ByteValue bitmask
DoubleValue hubScore
DoubleValue authorityScore
DoubleValue pageRankScore
FloatValue adamicAdarScore
Vertex<K,V> outVertex
Tuple2<T0,T1> outMsg
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
org.apache.flink.graph.pregel.ComputeFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
boolean setNewVertexValueCalled
Object first
long numberOfVertices
boolean optDegrees
IterationRuntimeContext runtimeContext
Collector<T> out
Collector<T> outWithDegrees
Vertex<K,V> outVal
Vertex<K,V> outValWithDegrees
long inDegree
long outDegree
boolean setNewVertexValueCalled
long numberOfVertices
EdgeDirection direction
Tuple2<T0,T1> outValue
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
Object vertexId
org.apache.flink.graph.spargel.ScatterFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
long inDegree
long outDegree
boolean isBounded
byte[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.ByteValueArray.ReadIterator iterator
boolean ascendingComparison
ByteValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
char[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.CharValueArray.ReadIterator iterator
boolean ascendingComparison
CharValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
double[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.DoubleValueArray.ReadIterator iterator
boolean ascendingComparison
DoubleValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
float[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.FloatValueArray.ReadIterator iterator
boolean ascendingComparison
FloatValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
int[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.IntValueArray.ReadIterator iterator
boolean ascendingComparison
IntValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
long[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.LongValueArray.ReadIterator iterator
boolean ascendingComparison
LongValueArray reference
TypeComparator<T>[] comparators
int position
IntValue hashValue
org.apache.flink.graph.types.valuearray.NullValueArray.ReadIterator iterator
boolean ascendingComparison
NullValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
short[] data
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.ShortValueArray.ReadIterator iterator
boolean ascendingComparison
ShortValueArray reference
TypeComparator<T>[] comparators
boolean isBounded
int boundedLength
byte[] data
int length
int position
MurmurHash hash
IntValue hashValue
org.apache.flink.graph.types.valuearray.StringValueArray.ReadIterator iterator
boolean ascendingComparison
StringValueArray reference
TypeComparator<T>[] comparators
TypeInformation<T> valueType
Class<T> type
Object value
int seed
int count
int hash
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
TypeSerializer<T> keySerializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
Configuration configuration
org.apache.hive.hcatalog.mapreduce.HCatInputFormat hCatInputFormat
org.apache.hadoop.mapreduce.RecordReader<KEYIN,VALUEIN> recordReader
boolean fetched
boolean hasNext
String[] fieldNames
org.apache.hive.hcatalog.data.schema.HCatSchema outputSchema
TypeInformation<T> resultType
MesosWorkerStore.Worker worker
org.apache.mesos.Protos.TaskID taskID
ResourceProfile profile
scala.Option<A> slaveID
scala.Option<A> hostname
MesosWorkerStore.WorkerState state
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
scala.collection.Seq<A> tasks
scala.collection.Seq<A> newLeases
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
String hostname
private Object readResolve()
private Object readResolve()
scala.collection.Seq<A> tasks
boolean replace
private Object readResolve()
scala.collection.immutable.Map<A,B> remaining
int retries
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.SlaveID slaveID
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
private Object readResolve()
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.SlaveID slaveID
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.mesos.scheduler.TaskMonitor.TaskGoalState goal
private Object readResolve()
private Object readResolve()
org.apache.flink.mesos.scheduler.TaskMonitor.TaskGoalState state
private Object readResolve()
org.apache.mesos.Protos.TaskID taskID
org.apache.mesos.Protos.TaskStatus status
private Object readResolve()
String hostname
Collection<E> offerIds
Collection<E> operations
org.apache.mesos.Protos.Filters filters
String message
org.apache.mesos.Protos.ExecutorID executorId
org.apache.mesos.Protos.SlaveID slaveId
int status
org.apache.mesos.Protos.ExecutorID executorId
org.apache.mesos.Protos.SlaveID slaveId
byte[] data
org.apache.mesos.Protos.OfferID offerId
org.apache.mesos.Protos.FrameworkID frameworkId
org.apache.mesos.Protos.MasterInfo masterInfo
org.apache.mesos.Protos.MasterInfo masterInfo
org.apache.mesos.Protos.SlaveID slaveId
org.apache.mesos.Protos.TaskStatus status
int fieldNumber
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
int batchSize
Configuration conf
org.apache.orc.TypeDescription schema
int[] selectedFields
ArrayList<E> conjunctPredicates
Serializable lowerBound
Serializable upperBound
Serializable[] literals
OrcRowInputFormat.Predicate pred
OrcRowInputFormat.Predicate[] preds
PythonDualInputStreamer<IN1,IN2,OUT> streamer
PythonSingleInputStreamer<IN,OUT> streamer
long mappedFileSizeBytes
boolean readAsByteArray
long mappedFileSizeBytes
Configuration config
Configuration config
int envID
int setID
int port
PythonSender sender
PythonReceiver<OUT> receiver
AtomicReference<V> msg
AbstractRichFunction function
JobID jobID
ExecutionAttemptID executionAttemptID
SerializedValue<T> userAccumulators
byte[] key
org.apache.flink.runtime.blob.BlobKey.BlobType type
AbstractID random
long checkpointId
long triggerTimestamp
Map<K,V> taskStats
TaskStateStats
accessible by their ID.int numberOfSubtasks
CheckpointProperties props
long checkpointId
long timestamp
long bytesBufferedInAlignment
long alignmentDurationNanos
long syncDurationMillis
long asyncDurationMillis
CheckpointType checkpointType
CheckpointStorageLocationReference targetLocation
CheckpointType checkpointType
boolean forced
boolean discardSubsumed
boolean discardFinished
boolean discardCancelled
boolean discardFailed
boolean discardSuspended
long numRestoredCheckpoints
long numTotalCheckpoints
int numInProgressCheckpoints
long numCompletedCheckpoints
long numFailedCheckpoints
List<E> checkpointsHistory
CheckpointStatsHistory.createSnapshot()
.Map<K,V> checkpointsById
CheckpointStatsHistory.createSnapshot()
.int maxSize
boolean readOnly
CompletedCheckpointStats latestCompletedCheckpoint
FailedCheckpointStats latestFailedCheckpoint
CompletedCheckpointStats latestSavepoint
CheckpointStatsCounts counts
CompletedCheckpointStatsSummary summary
CheckpointStatsHistory history
RestoredCheckpointStats latestRestoredCheckpoint
CheckpointDeclineReason checkpointDeclineReason
JobID job
long checkpointID
long timestamp
long duration
Map<K,V> operatorStates
CheckpointProperties props
Collection<E> masterHookStates
CompletedCheckpointStorageLocation storageLocation
StreamStateHandle metadataHandle
String externalPointer
long stateSize
long alignmentBuffered
SubtaskStateStats latestAcknowledgedSubtask
String externalPointer
boolean discarded
MinMaxAvgStats stateSize
MinMaxAvgStats duration
MinMaxAvgStats alignmentBuffered
boolean isDiscarded
boolean registered
boolean discarded
int numAcknowledgedSubtasks
long stateSize
long alignmentBuffered
long failureTimestamp
SubtaskStateStats latestAcknowledgedSubtask
String failureMsg
long restoreCheckpointId
TaskStateSnapshot taskStateSnapshot
String name
byte[] bytes
int version
long min
long max
long sum
long count
OperatorID operatorID
Map<K,V> operatorSubtaskStates
int parallelism
int maxParallelism
StateObjectCollection<T extends StateObject> managedOperatorState
OperatorStateBackend
.StateObjectCollection<T extends StateObject> rawOperatorState
OperatorStateCheckpointOutputStream
.StateObjectCollection<T extends StateObject> managedKeyedState
KeyedStateBackend
.StateObjectCollection<T extends StateObject> rawKeyedState
KeyedStateCheckpointOutputStream
.long stateSize
int currentNumAcknowledgedSubtasks
long currentStateSize
long currentAlignmentBuffered
SubtaskStateStats latestAcknowledgedSubtask
long checkpointId
CheckpointProperties props
long restoreTimestamp
String externalPath
Collection<E> stateObjects
ChainedStateHandle<T extends StateObject> managedOperatorState
OperatorStateBackend
.ChainedStateHandle<T extends StateObject> rawOperatorState
OperatorStateCheckpointOutputStream
.KeyedStateHandle managedKeyedState
KeyedStateBackend
.KeyedStateHandle rawKeyedState
KeyedStateCheckpointOutputStream
.long stateSize
int subtaskIndex
long ackTimestamp
long stateSize
long syncCheckpointDuration
long asyncCheckpointDuration
long alignmentBuffered
long alignmentDuration
JobVertexID jobVertexID
Map<K,V> subtaskStates
int parallelism
int maxParallelism
int chainLength
JobVertexID jobVertexId
SubtaskStateStats[] subtaskStats
TaskStateStats.TaskStateStatsSummary summaryStats
int numAcknowledgedSubtasks
SubtaskStateStats latestAckedSubtaskStats
MinMaxAvgStats stateSize
MinMaxAvgStats ackTimestamp
MinMaxAvgStats syncCheckpointDuration
MinMaxAvgStats asyncCheckpointDuration
MinMaxAvgStats alignmentBuffered
MinMaxAvgStats alignmentDuration
JobID jobID
long totalContainerMemoryMB
long taskManagerHeapSizeMB
long taskManagerDirectMemoryLimitMB
int numSlots
HashMap<K,V> taskManagerEnv
Configuration systemProperties
List<E> artifacts
Map<K,V> environmentVariables
Configuration dynamicConfiguration
private Object readResolve() throws ObjectStreamException
ObjectStreamException
String message
SerializedThrowable error
private Object readResolve() throws ObjectStreamException
ObjectStreamException
int numRegisteredTaskManagers
int totalNumberOfSlots
ResourceID resourceID
akka.actor.ActorRef resourceManager
long connectionId
private Object readResolve() throws ObjectStreamException
ObjectStreamException
private Object readResolve() throws ObjectStreamException
ObjectStreamException
akka.actor.ActorRef resourceManager
akka.actor.ActorRef jobManager
Collection<E> currentlyRegisteredTaskManagers
ResourceID resourceId
ResourceID resourceId
String message
int numberOfWorkers
JobID jobId
ApplicationStatus finalStatus
String message
String jobManagerAddress
private Object readResolve() throws ObjectStreamException
ObjectStreamException
String resourceId
double cpuCores
int heapMemoryInMB
int directMemoryInMB
int nativeMemoryInMB
int networkMemoryInMB
Map<K,V> extendedResources
ResourceSpec
.ResourceID resourceId
int slotNumber
ResultPartitionID consumedPartitionId
ResultPartitionLocation consumedPartitionLocation
IntermediateDataSetID consumedResultId
ResultPartitionType consumedPartitionType
int consumedSubpartitionIndex
DistributionPattern
and the subtask indices of the producing and consuming task.InputChannelDeploymentDescriptor[] inputChannels
IntermediateDataSetID resultId
IntermediateResultPartitionID partitionId
ResultPartitionType partitionType
int numberOfSubpartitions
int maxParallelism
boolean sendScheduleOrUpdateConsumersMessage
org.apache.flink.runtime.deployment.ResultPartitionLocation.LocationType locationType
ConnectionID connectionId
TaskDeploymentDescriptor.MaybeOffloaded<T> serializedJobInformation
TaskDeploymentDescriptor.MaybeOffloaded<T> serializedTaskInformation
JobID jobId
NOTE: this is redundant to the information stored in TaskDeploymentDescriptor.serializedJobInformation
but
needed in order to restore offloaded data.
ExecutionAttemptID executionId
AllocationID allocationId
int subtaskIndex
int attemptNumber
Collection<E> producedPartitions
Collection<E> inputGates
int targetSlotNumber
JobManagerTaskRestore taskRestore
SerializedValue<T> serializedValue
PermanentBlobKey serializedValueKey
String blobServerHostname
int blobServerPort
ExecutionAttemptID attemptId
long[] stateTimestamps
int attemptNumber
ExecutionState state
String failureCause
TaskManagerLocation assignedResourceLocation
AllocationID assignedAllocationID
StringifiedAccumulatorResult[] userAccumulators
int parallelSubtaskIndex
IOMetrics ioMetrics
JobID jobID
String jobName
Map<K,V> tasks
List<E> verticesInCreationOrder
long[] stateTimestamps
System.currentTimeMillis()
when
the execution graph transitioned into a certain state. The index into this array is the
ordinal of the enum value, i.e. the timestamp when the graph went into state "RUNNING" is
at stateTimestamps[RUNNING.ordinal()]
.JobStatus state
ErrorInfo failureCause
String jsonPlan
StringifiedAccumulatorResult[] archivedUserAccumulators
ArchivedExecutionConfig archivedExecutionConfig
boolean isStoppable
Map<K,V> serializedUserAccumulators
CheckpointCoordinatorConfiguration jobCheckpointingConfiguration
CheckpointStatsSnapshot checkpointStatsSnapshot
ArchivedExecutionVertex[] taskVertices
JobVertexID id
String name
int parallelism
int maxParallelism
StringifiedAccumulatorResult[] archivedUserAccumulators
int subTaskIndex
EvictingBoundedList<T> priorExecutions
String taskNameWithSubtask
ArchivedExecution currentExecution
SerializedThrowable exception
long timestamp
long expectedModVersion
long actualModVersion
long numRecordsIn
long numRecordsOut
double numRecordsInPerSecond
double numRecordsOutPerSecond
long numBytesInLocal
long numBytesInRemote
long numBytesOut
double numBytesInLocalPerSecond
double numBytesInRemotePerSecond
double numBytesOutPerSecond
JobID jobId
String jobName
SerializedValue<T> serializedExecutionConfig
Configuration jobConfiguration
Collection<E> requiredJarFileBlobKeys
Collection<E> requiredClasspathURLs
IntermediateDataSetID intermediateDataSetID
InputChannelDeploymentDescriptor inputChannelDeploymentDescriptor
JobVertexID jobVertexId
String taskName
int numberOfSubtasks
int maxNumberOfSubtaks
String invokableClassName
Configuration taskConfiguration
int maxAttempts
long delay
akka.actor.ActorRef actor
UUID leaderSessionID
MessageDecorator decorator
scala.concurrent.ExecutionContext executionContext
String path
int numberOfCPUCores
long sizeOfPhysicalMemory
long sizeOfJvmHeap
long sizeOfManagedMemory
Instance instance
boolean closed
InetSocketAddress address
int connectionIndex
SocketAddress address
ResultPartitionID partitionId
IntermediateResultPartitionID partitionId
ExecutionAttemptID producerId
String formatDescription
IntermediateDataSetID id
JobVertex producer
List<E> consumers
ResultPartitionType resultType
JobVertex target
DistributionPattern distributionPattern
IntermediateDataSet source
IntermediateDataSetID sourceId
String shipStrategyName
String preProcessingOperationName
String operatorLevelCachingDescription
Map<K,V> taskVertices
Configuration jobConfiguration
JobID jobID
String jobName
long sessionTimeout
boolean allowQueuedScheduling
ScheduleMode scheduleMode
SerializedValue<T> serializedExecutionConfig
JobCheckpointingSettings snapshotSettings
SavepointRestoreSettings savepointRestoreSettings
List<E> userJars
Map<K,V> userArtifacts
List<E> userJarBlobKeys
List<E> classpaths
JobVertexID id
ArrayList<E> idAlternatives
ArrayList<E> operatorIDs
ArrayList<E> operatorIdsAlternatives
ArrayList<E> results
ArrayList<E> inputs
int parallelism
int maxParallelism
ResourceSpec minResources
ResourceSpec preferredResources
Configuration configuration
String invokableClassName
boolean isStoppable
InputSplitSource<T extends InputSplit> inputSplitSource
String name
SlotSharingGroup slotSharingGroup
CoLocationGroup coLocationGroup
String operatorName
String operatorDescription
String operatorPrettyName
String resultOptimizerProperties
InputDependencyConstraint inputDependencyConstraint
String formatDescription
String restorePath
boolean allowNonRestoredState
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
CheckpointRetentionPolicy checkpointRetentionPolicy
boolean isExactlyOnce
false
, at least once mode has been configured. This is
not a necessary attribute, because the checkpointing mode is only relevant
for the stream tasks, but we expose it here to forward it to the web runtime
UI.List<E> verticesToTrigger
List<E> verticesToAcknowledge
List<E> verticesToConfirm
CheckpointCoordinatorConfiguration checkpointCoordinatorConfiguration
SerializedValue<T> defaultStateBackend
SerializedValue<T> masterHooks
AbstractID id
List<E> vertices
Set<E> ids
SlotSharingGroupId slotSharingGroupId
int slotIndex
AllocationID allocationId
JobID jobId
Collection<E> allocatedSlotInfos
ResourceID resourceID
ResourceManagerId resourceManagerId
ResourceID resourceManagerResourceId
JobID jobId
ApplicationStatus applicationStatus
Map<K,V> accumulatorResults
long netRuntime
SerializedThrowable serializedThrowable
null
if the job finished successfully.byte[] inputSplitData
int blobManagerPort
Collection<E> requiredJarFiles
Collection<E> requiredClasspaths
protected Object readResolve() throws ObjectStreamException
ObjectStreamException
scala.Option<A> job
private Object readResolve()
scala.collection.Iterable<A> jobs
private Object readResolve()
JobID jobID
ArchivedExecutionGraph graph
private Object readResolve()
JobID jobID
private Object readResolve()
private Object readResolve()
private Object readResolve()
JobID jobID
JobVertexID vertexID
String taskName
int totalNumberOfSubTasks
int subtaskIndex
ExecutionAttemptID executionID
ExecutionState newExecutionState
long timestamp
String optionalMessage
private Object readResolve()
private Object readResolve()
JobID jobID
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
Integer blobManagerPort
Collection<E> requiredJarFiles
Collection<E> requiredClasspaths
private Object readResolve()
private Object readResolve()
String savepointPath
private Object readResolve()
Throwable cause
private Object readResolve()
private Object readResolve()
scala.Option<A> leaderSessionID
private Object readResolve()
JobID jobID
AccessExecutionGraph executionGraph
private Object readResolve()
private Object readResolve()
JobID jobID
private Object readResolve()
SerializedThrowable cause
private Object readResolve()
SerializedJobExecutionResult result
private Object readResolve()
JobID jobId
private Object readResolve()
private Object readResolve()
byte[] splitData
private Object readResolve()
private Object readResolve()
JobID jobId
private Object readResolve()
SubmittedJobGraph submittedJobGraph
private Object readResolve()
scala.collection.Iterable<A> taskManagers
private Object readResolve()
JobID jobID
ListeningBehaviour listeningBehaviour
private Object readResolve()
JobID jobId
private Object readResolve()
JobID jobID
private Object readResolve()
JobID jobID
boolean removeJobFromStateBackend
private Object readResolve()
private Object readResolve()
private Object readResolve()
JobID jobId
private Object readResolve()
JobID jobID
private Object readResolve()
private Object readResolve()
JobID jobID
private Object readResolve()
private Object readResolve()
JobID jobID
JobVertexID vertexID
ExecutionAttemptID executionAttempt
private Object readResolve()
private Object readResolve()
JobID jobId
IntermediateDataSetID intermediateDataSetId
ResultPartitionID resultPartitionId
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
InstanceID instanceID
private Object readResolve()
ResourceID resourceId
private Object readResolve()
private Object readResolve()
akka.actor.ActorRef actor
private Object readResolve()
UUID leaderSessionID
private Object readResolve()
private Object readResolve()
scala.collection.Iterable<A> runningJobs
private Object readResolve()
scala.collection.Iterable<A> runningJobs
private Object readResolve()
JobID jobId
ResultPartitionID partitionId
private Object readResolve()
JobID jobID
private Object readResolve()
private Object readResolve()
JobID jobID
private Object readResolve()
JobGraph jobGraph
ListeningBehaviour listeningBehaviour
private Object readResolve()
scala.Option<A> instance
private Object readResolve()
JobID jobId
scala.Option<A> savepointDirectory
private Object readResolve()
private Object readResolve()
private Object readResolve()
UUID leaderSessionID
InstanceID instanceId
Exception cause
private Object readResolve()
InstanceID instanceID
int blobPort
private Object readResolve()
InstanceID instanceID
int blobPort
private Object readResolve()
Throwable reason
private Object readResolve()
ResourceID resourceId
TaskManagerLocation connectionInfo
HardwareDescription resources
int numberOfSlots
private Object readResolve()
private Object readResolve()
InstanceID instanceId
String stackTrace
int sampleId
ExecutionAttemptID executionId
Time delayBetweenSamples
int maxStackTraceDepth
int numRemainingSamples
List<E> currentTraces
akka.actor.ActorRef sender
private Object readResolve()
int sampleId
ExecutionAttemptID executionId
int numSamples
Time delayBetweenSamples
int maxStackTraceDepth
private Object readResolve()
int sampleId
ExecutionAttemptID executionAttemptID
List<E> samples
private Object readResolve()
InstanceID instanceID
scala.collection.Seq<A> accumulators
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
UUID leaderId
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.runtime.messages.TaskManagerMessages.LogTypeRequest requestType
private Object readResolve()
int number
private Object readResolve()
int number
private Object readResolve()
private Object readResolve()
private Object readResolve()
InstanceID instanceID
String stackTrace
private Object readResolve()
private Object readResolve()
ExecutionAttemptID attemptID
private Object readResolve()
ExecutionAttemptID executionID
private Object readResolve()
ExecutionAttemptID executionID
Throwable cause
private Object readResolve()
ExecutionAttemptID attemptID
private Object readResolve()
TaskDeploymentDescriptor tasks
private Object readResolve()
ExecutionAttemptID executionID
private Object readResolve()
TaskExecutionState taskExecutionState
private Object readResolve()
ExecutionAttemptID executionID
Iterable<T> partitionInfos
private Object readResolve()
ExecutionAttemptID executionID
IntermediateDataSetID resultId
InputChannelDeploymentDescriptor partitionInfo
private Object readResolve()
JobID job
ExecutionAttemptID taskExecutionId
long checkpointId
TaskStateSnapshot subtaskState
CheckpointMetrics checkpointMetrics
Throwable reason
long timestamp
long timestamp
CheckpointOptions checkpointOptions
int numTaskManagersConnected
int numSlotsTotal
int numSlotsAvailable
Collection<E> jobsWithStatus
int numJobsRunningOrPending
int numJobsFinished
int numJobsCancelled
int numJobsFailed
Collection<E> jobs
boolean includeRunning
boolean includeFinished
private Object readResolve()
private Object readResolve()
private Object readResolve()
byte[] serializedCounters
byte[] serializedGauges
byte[] serializedMeters
byte[] serializedHistograms
int numCounters
int numGauges
int numMeters
int numHistograms
int cnt
int cnt
Record res
int cnt
IntValue theInteger
int cnt
int cnt
int cnt
StringBuilder bld
int cnt
int cnt
int cnt
Integer count
boolean opened
boolean closed
boolean opened
boolean closed
int val
int size
int val
boolean isLong
int val
TypeComparatorFactory<T> typeComparator
int parallelism
TypeComparatorFactory<T> comparatorFactory
Configuration config
JobID jobId
JobVertexID jobVertexId
int numKeyGroups
String registrationName
KvStateID[] kvStateIds
InetSocketAddress[] kvStateAddresses
int numRegisteredKeyGroups
JobID jobId
JobVertexID jobVertexId
KeyGroupRange keyGroupRange
String registrationName
KvStateID kvStateId
InetSocketAddress kvStateServerAddress
JobID jobId
JobVertexID jobVertexId
KeyGroupRange keyGroupRange
String registrationName
String reason
int numberTaskManagers
int numberRegisteredSlots
int numberFreeSlots
JobID jobId
AllocationID allocationId
ResourceProfile resourceProfile
String targetAddress
int responseCode
int sampleId
long endTimestamp
double[] subTaskBackPressureRatio
double maxSubTaskBackPressureRatio
OperatorBackPressureStats operatorBackPressureStats
boolean numBytesInLocalComplete
boolean numBytesInRemoteComplete
boolean numBytesOutComplete
boolean numRecordsInComplete
boolean numRecordsOutComplete
JobIDDeserializer jobIdDeserializer
SerializedThrowableDeserializer serializedThrowableDeserializer
SerializedValueDeserializer serializedValueDeserializer
JobIDSerializer jobIdSerializer
SerializedValueSerializer serializedValueSerializer
SerializedThrowableSerializer serializedThrowableSerializer
TaskManagerMetricsInfo taskManagerMetrics
ResourceID resourceId
String address
int dataPort
long lastHeartbeat
int numberSlots
int numberAvailableSlots
HardwareDescription hardwareDescription
int responseCode
Serializable fencingToken
Serializable payload
private void readObject(ObjectInputStream ois) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
SerializedValue<T> serializedMethodInvocation
TypeSerializer<T> elementSerializer
protected final Object readResolve() throws ObjectStreamException
ObjectStreamException
byte[] encodedReference
DirectoryStateHandle directoryStateHandle
KeyGroupRange keyGroupRange
Path directory
long checkpointId
UUID backendIdentifier
StreamStateHandle metaDataState
Set<E> sharedStateHandleIDs
UUID backendIdentifier
SharedStateRegistry
.KeyGroupRange keyGroupRange
long checkpointId
Map<K,V> sharedState
Map<K,V> privateState
StreamStateHandle metaStateHandle
int startKeyGroup
int endKeyGroup
KeyGroupRange keyGroupRange
long[] offsets
KeyGroupRangeOffsets groupRangeOffsets
StreamStateHandle stateHandle
File[] allocationBaseDirs
JobID jobID
JobVertexID jobVertexID
int subtaskIndex
long[] offsets
OperatorStateHandle.Mode distributionMode
Map<K,V> stateNameToPartitionOffsets
StreamStateHandle delegateStateHandle
StreamStateHandle wrappedStreamStateHandle
StateObject jobManagerOwnedSnapshot
StateObject taskLocalSnapshot
Path filePath
long stateSize
Path exclusiveCheckpointDir
FileStateHandle metadataFileHandle
String externalPointer
int fileStateThreshold
TernaryBoolean asynchronousSnapshots
boolean disabled
byte[] data
String handleName
ByteStreamStateHandle
with the exact same name must also have the exact same content in data.int maxStateSize
TernaryBoolean asynchronousSnapshots
Object userValue
long lastAccessTimestamp
Collection<E> accumulatorSnapshots
Collection<E> slotsStatus
SlotID slotID
ResourceProfile resourceProfile
AllocationID allocationID
JobID jobID
InstanceID registrationId
ResourceID resourceManagerResourceId
ClusterInformation clusterInformation
AllocationID allocationId
JobID jobId
AllocationID allocationId
int slotIndex
ResourceProfile resourceProfile
JobID jobID
ExecutionAttemptID executionId
ExecutionState executionState
SerializedThrowable throwable
AccumulatorSnapshot accumulators
IOMetrics ioMetrics
ResourceID resourceID
InetAddress inetAddress
String fqdnHostName
String hostName
int dataPort
String stringRepresentation
UUID leaderSessionID
UUID leaderSessionID
akka.actor.ActorRef testActor
int numKeys
int offsetSeconds
int sleepMs
int durationMs
long ms
int saveRecordCnt
int lostRecordCnt
CheckpointingMode checkpointingMode
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
boolean forceCheckpointing
CheckpointConfig.ExternalizedCheckpointCleanup externalizedCheckpointCleanup
boolean failOnCheckpointingErrors
long maxTimestamp
Comparator comparator
boolean byAggregate
boolean first
FieldAccessor<T,F> fieldAccessor
FieldAccessor<T,F> fieldAccessor
SumFunction adder
TypeSerializer<T> serializer
boolean isTuple
OutputFormat<IT> format
boolean cleanupCalled
PrintSinkOutputWriter<IN> writer
SerializableObject lock
SerializationSchema<T> schema
String hostName
int port
int maxNumRetries
boolean autoFlush
int retries
boolean isRunning
LinkedHashMap<K,V> pendingCommitTransactions
java.time.Clock clock
ListStateDescriptor<T> stateDescriptor
TwoPhaseCommitSinkFunction.TransactionHolder<TXN> currentTransactionHolder
long transactionTimeout
boolean ignoreFailuresAfterTransactionTimeout
TwoPhaseCommitSinkFunction.recoverAndCommit(Object)
will be caught instead of
propagated.double transactionTimeoutWarningRatio
TypeSerializer<T> transactionSerializer
TypeSerializer<T> contextSerializer
String path
ArrayList<E> tupleList
WriteFormat<IN> format
long millis
long lastTime
long bucketCheckInterval
StreamingFileSink.BucketsBuilder<IN,BucketID> bucketsBuilder
long bucketCheckInterval
Path basePath
BulkWriter.Factory<T> writerFactory
BucketAssigner<IN,BucketID> bucketAssigner
org.apache.flink.streaming.api.functions.sink.filesystem.BucketFactory<IN,BucketID> bucketFactory
long bucketCheckInterval
Path basePath
Encoder<IN> encoder
BucketAssigner<IN,BucketID> bucketAssigner
RollingPolicy<IN,BucketID> rollingPolicy
org.apache.flink.streaming.api.functions.sink.filesystem.BucketFactory<IN,BucketID> bucketFactory
String formatString
java.time.ZoneId zoneId
long partSize
long rolloverInterval
long inactivityInterval
String path
int readerParallelism
FileInputFormat<OT> format
FileInputFormat
to be read.long interval
FileProcessingMode watchType
FileProcessingMode
.long globalModificationTime
boolean isRunning
FileInputFormat<OT> format
TypeSerializer<T> serializer
TypeSerializer<T> serializer
byte[] elementsSerialized
int numElements
int numElementsEmitted
int numElementsToSkip
boolean isRunning
SplittableIterator<T> fullIterator
boolean isRunning
TypeInformation<T> typeInfo
InputFormat<OT,T extends InputSplit> format
boolean isRunning
TypeSerializer<T> idSerializer
long start
long end
boolean isRunning
long modificationTime
Serializable splitState
long currentTimestamp
AscendingTimestampExtractor.MonotonyViolationHandler violationHandler
long currentMaxTimestamp
long lastEmittedWatermark
long maxOutOfOrderness
AggregateFunction<IN,ACC,OUT> aggFunction
AggregateFunction<IN,ACC,OUT> aggFunction
FoldFunction<O,T> foldFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
FoldFunction<O,T> foldFunction
ProcessAllWindowFunction<IN,OUT,W extends Window> windowFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
TypeInformation<T> accTypeInformation
FoldFunction<O,T> foldFunction
ProcessWindowFunction<IN,OUT,KEY,W extends Window> windowFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
TypeInformation<T> accTypeInformation
FoldFunction<O,T> foldFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
Window window
ProcessWindowFunction.Context context
ReduceFunction<T> reduceFunction
ReduceFunction<T> reduceFunction
ProcessAllWindowFunction<IN,OUT,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
ProcessWindowFunction<IN,OUT,KEY,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
int[] order
int fieldId
int fieldId
int[] indexes
Configuration config
String edgeId
int sourceId
int targetId
int typeNumber
List<E> selectedNames
OutputTag<T> outputTag
StreamEdge
.StreamPartitioner<T> outputPartitioner
StreamPartitioner
on this StreamEdge
.String sourceOperatorName
String targetOperatorName
int id
Integer parallelism
int maxParallelism
ResourceSpec minResources
ResourceSpec preferredResources
Long bufferTimeout
String operatorName
String slotSharingGroup
String coLocationGroup
KeySelector<IN,KEY> statePartitioner1
KeySelector<IN,KEY> statePartitioner2
TypeSerializer<T> stateKeySerializer
List<E> outputSelectors
TypeSerializer<T> typeSerializerIn1
TypeSerializer<T> typeSerializerIn2
TypeSerializer<T> typeSerializerOut
List<E> inEdges
List<E> outEdges
Class<T> jobVertexClass
InputFormat<OT,T extends InputSplit> inputFormat
String transformationUID
String userHash
ChainingStrategy chainingStrategy
long combinedWatermark
long input1Watermark
long input2Watermark
Function userFunction
TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
long currentWatermark
InternalTimerService
.byte[] serializedInitialValue
TypeSerializer<T> outTypeSerializer
TypeSerializer<T> serializer
TypeSerializer<T> outSerializer
int[] fields
int numFields
long currentWatermark
InternalTimerService
.TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
int length
boolean immutableType
int capacity
AsyncDataStream.OutputMode outputMode
long timeout
long currentWatermark
InternalTimerService
.List<E> broadcastStateDescriptors
long currentWatermark
InternalTimerService
.long lowerBound
long upperBound
TypeSerializer<T> leftTypeSerializer
TypeSerializer<T> rightTypeSerializer
long size
SessionWindowTimeGapExtractor<T> sessionWindowTimeGapExtractor
SessionWindowTimeGapExtractor<T> sessionWindowTimeGapExtractor
long sessionTimeout
long sessionTimeout
long size
long slide
long offset
long size
long offset
long slide
long size
long offset
long size
long offset
long maxCount
boolean doEvictAfter
DeltaFunction<DATA> deltaFunction
double threshold
boolean doEvictAfter
long windowSize
boolean doEvictAfter
long interval
ReducingStateDescriptor<T> stateDesc
long interval
ReducingStateDescriptor<T> stateDesc
long maxCount
ReducingStateDescriptor<T> stateDesc
DeltaFunction<DATA> deltaFunction
double threshold
ValueStateDescriptor<T> stateDesc
String insertQuery
ClusterBuilder builder
String keySpace
String table
Map<K,V> lastCommittedCheckpoints
CassandraCommitter.isCheckpointCommitted(int, long)
.Class<T> clazz
MapperOptions options
String keyspace
int rowArity
String insertQuery
ClusterBuilder builder
int arity
org.slf4j.Logger log
AtomicReference<V> throwable
com.google.common.util.concurrent.FutureCallback<V> callback
Semaphore semaphore
ClusterBuilder builder
CassandraSinkBaseConfig config
CassandraFailureHandler failureHandler
int maxConcurrentRequests
java.time.Duration maxConcurrentRequestsTimeout
String insertQuery
ClusterBuilder builder
Integer bulkProcessorFlushMaxActions
Integer bulkProcessorFlushMaxSizeMb
Long bulkProcessorFlushIntervalMillis
ElasticsearchSinkBase.BulkFlushBackoffPolicy bulkProcessorFlushBackoffPolicy
Map<K,V> userConfig
For TransportClient
based implementations, this config
map would also contain Elasticsearch-shipped configuration, and therefore this config map
would also be forwarded when creating the Elasticsearch client.
ElasticsearchSinkFunction<T> elasticsearchSinkFunction
ActionRequests
from each incoming element.ActionRequestFailureHandler failureHandler
ActionRequests
.boolean flushOnCheckpoint
ElasticsearchApiCallBridge<C extends AutoCloseable> callBridge
AtomicLong numPendingRequests
ElasticsearchSinkBase.flushOnCheckpoint
is true
.
This is incremented whenever the user adds (or re-adds through the ActionRequestFailureHandler
) requests
to the RequestIndexer
. It is decremented for each completed request of a bulk request, in
BulkProcessor.Listener#afterBulk(long, BulkRequest, BulkResponse)
and
BulkProcessor.Listener#afterBulk(long, BulkRequest, Throwable)
.
AtomicReference<V> failureThrowable
BulkProcessor.Listener
if a Throwable
was thrown in callbacks and
the user considered it should fail the sink via the
ActionRequestFailureHandler.onFailure(ActionRequest, Throwable, int, RequestIndexer)
method.
Errors will be checked and rethrown before processing each input element, and when the sink is closed.
ElasticsearchSinkBase.FlushBackoffType backoffType
int maxRetryCount
long delayMillis
String index
String docType
String keyDelimiter
String keyNullLiteral
SerializationSchema<T> serializationSchema
org.elasticsearch.common.xcontent.XContentType contentType
ElasticsearchUpsertTableSinkBase.RequestFactory requestFactory
int[] keyFieldIndices
List<E> transportAddresses
We are using InetSocketAddress
because TransportAddress
is not serializable in Elasticsearch 2.x.
List<E> transportAddresses
We are using InetSocketAddress
because TransportAddress
is not serializable in Elasticsearch 5.x.
List<E> httpHosts
RestClientFactory restClientFactory
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
String formatString
String basePath
Path
that stores all bucket directories.Bucketer bucketer
Bucketer
that is used to determine the path of bucket directories.Writer<T> writerTemplate
Writer<T> writer
long batchSize
String inProgressSuffix
String inProgressPrefix
String pendingSuffix
String pendingPrefix
String validLengthSuffix
String validLengthPrefix
String partPrefix
long asyncTimeout
Configuration fsConfig
String currentFile
long currentFileValidLength
List<E> pendingFiles
Map<K,V> pendingFilesPerCheckpoint
boolean syncOnFlush
String charsetName
String basePath
Path
that stores all bucket directories.Bucketer<T> bucketer
Bucketer
that is used to determine the path of bucket directories.Writer<T> writerTemplate
long batchSize
long inactiveBucketCheckInterval
long inactiveBucketThreshold
long batchRolloverInterval
String inProgressSuffix
String inProgressPrefix
String pendingSuffix
String pendingPrefix
String validLengthSuffix
String validLengthPrefix
String partPrefix
String partSuffix
boolean useTruncate
long asyncTimeout
ms
).Configuration fsConfig
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
String formatString
java.time.ZoneId zoneId
FlinkKafka011ErrorCode errorCode
Properties properties
long pollTimeout
Properties kafkaProperties
Properties properties
long pollTimeout
FlinkConnectorRateLimiter rateLimiter
FlinkKafkaConsumer09.setRateLimiter(FlinkConnectorRateLimiter)
.KafkaTopicsDescriptor topicsDescriptor
KafkaDeserializationSchema<T> deserializer
Map<K,V> subscribedPartitionsToStartOffsets
SerializedValue<T> periodicWatermarkAssigner
SerializedValue<T> punctuatedWatermarkAssigner
boolean enableCommitOnCheckpoints
boolean filterRestoredPartitionsWithCurrentTopicsDescriptor
OffsetCommitMode offsetCommitMode
FlinkKafkaConsumerBase.open(Configuration)
since it depends
on whether or not checkpointing is enabled for the job.long discoveryIntervalMillis
StartupMode startupMode
StartupMode.GROUP_OFFSETS
).Map<K,V> specificStartupOffsets
StartupMode.SPECIFIC_OFFSETS
.Long startupOffsetsTimestamp
StartupMode.TIMESTAMP
.org.apache.commons.collections.map.LinkedMap pendingOffsetsToCommit
boolean restoredFromOldState
boolean running
boolean useMetrics
true
, offset metrics (e.g. current offset, committed offset) and
Kafka-shipped metrics will be registered.FlinkKafkaErrorCode errorCode
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> schema
FlinkKafkaPartitioner<T> flinkKafkaPartitioner
Map<K,V> topicPartitionsMap
int kafkaProducersPoolSize
BlockingDeque<E> availableTransactionalIds
boolean writeTimestampToKafka
boolean logFailuresOnly
FlinkKafkaProducer.Semantic semantic
AtomicLong pendingRecords
Map<K,V> previouslyCreatedMetrics
boolean writeTimestampToKafka
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> schema
FlinkKafkaPartitioner<T> flinkKafkaPartitioner
Map<K,V> topicPartitionsMap
int kafkaProducersPoolSize
BlockingDeque<E> availableTransactionalIds
boolean writeTimestampToKafka
boolean logFailuresOnly
FlinkKafkaProducer011.Semantic semantic
AtomicLong pendingRecords
Map<K,V> previouslyCreatedMetrics
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> schema
FlinkKafkaPartitioner<T> flinkKafkaPartitioner
Map<K,V> topicPartitionsMap
boolean logFailuresOnly
boolean flushOnCheckpoint
SerializableObject pendingRecordsLock
long pendingRecords
DeserializationSchema<T> deserializationSchema
String topic
int partition
int cachedHash
int leaderId
int leaderPort
String leaderHost
KafkaTopicPartition topicPartition
int cachedHash
SerializationSchema<T> serializationSchema
int parallelInstanceId
KafkaPartitioner<T> kafkaPartitioner
int[] partitions
org.apache.nifi.remote.client.SiteToSiteClient client
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
NiFiDataPacketBuilder<T> builder
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
long waitTimeMs
boolean isRunning
String queueName
RMQConnectionConfig rmqConnectionConfig
SerializationSchema<T> schema
boolean logFailuresOnly
RMQSinkPublishOptions<IN> publishOptions
SerializableReturnListener returnListener
RMQConnectionConfig rmqConnectionConfig
String queueName
boolean usesCorrelationId
DeserializationSchema<T> schema
String host
Integer port
String virtualHost
String username
String password
String uri
Integer networkRecoveryInterval
Boolean automaticRecovery
Boolean topologyRecovery
Integer connectionTimeout
Integer requestedChannelMax
Integer requestedFrameMax
Integer requestedHeartbeat
Properties properties
TwitterSource.EndpointInitializer initializer
Random rnd
Random rnd
int counter
int counter
double errorProbability
int delayPerRecordMillis
boolean running
InetAddress hostIp
int port
TypeSerializer<T> serializer
long currentTimestamp
org.slf4j.Logger log
byte[] serFun
byte[] serFun
byte[] serFun
byte[] serFun
byte[] serFun
byte[] serSchema
String id
CheckpointCommitter committer
TypeSerializer<T> serializer
Set<E> pendingCheckpoints
long currentWatermark
WindowAssigner<T,W extends Window> windowAssigner
KeySelector<IN,KEY> keySelector
Trigger<T,W extends Window> trigger
StateDescriptor<S extends State,T> windowStateDescriptor
TypeSerializer<T> keySerializer
TypeSerializer<T> windowSerializer
long allowedLateness
window.maxTimestamp + allowedLateness
landmark.
OutputTag<T> lateDataOutputTag
OutputTag
to use for late arriving events. Elements for which
window.maxTimestamp + allowedLateness
is smaller than the current watermark will
be emitted to this.Window window
WindowOperator.AbstractPerWindowStateStore windowState
AggregateFunction<IN,ACC,OUT> aggFunction
AggregateFunction<IN,ACC,OUT> aggFunction
InternalProcessWindowContext<IN,OUT,KEY,W extends Window> ctx
Window window
InternalWindowFunction.InternalWindowContext internalContext
Partitioner<K> partitioner
KeySelector<IN,KEY> keySelector
KeySelector<IN,KEY> keySelector
int maxParallelism
int nextChannelToSendTo
int nextChannelToSendTo
Random random
int numberOfChannels
TypeSerializer<T> typeSerializer
int numKeys
int idlenessMs
int durationMs
long ms
long initialValue
ValueState<T> counter
ValueState<T> last
long numProcessedRecordsFailureThreshold
long numCompleteCheckpointsFailureThreshold
int maxNumFailures
long numProcessedRecords
long numCompleteCheckpoints
int sleepMs
int durationMs
int offsetSeconds
long ms
org.apache.flink.streaming.tests.SemanticsCheckMapper.ValidatorFunction validator
int payloadLength
int totalKeySpaceSize
long eventTimeClockProgressPerEvent
long maxOutOfOrder
long sleepTime
long sleepAfterElements
long monotonousEventTime
boolean running
TypeInformation<T> originalTypeInformation
int slideFactor
boolean failingTask
boolean killedJvm
int jvmPid
String taskNameWithSubtask
String allocationId
MapFunction<T,O> mapFunction
MapFunction<T,O> mapFunction
List<E> artificialStateBuilders
long eventTime
List<E> stringList
String strPayload
ComplexPayload.InnerPayLoad innerPayLoad
long sequenceNumber
AtomicReference<V> currentOwnerThread
ListStateDescriptor<T> listStateDescriptor
JoinFunction<IN1,IN2,OUT> keyedStateGenerator
JoinFunction<IN1,IN2,OUT> operatorStateGenerator
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
JoinFunction<IN1,IN2,OUT> stateValueGenerator
String stateName
ValueStateDescriptor<T> valueStateDescriptor
JoinFunction<IN1,IN2,OUT> stateValueGenerator
public void readExternal(ObjectInput in) throws IOException
IOException
public void writeExternal(ObjectOutput out) throws IOException
IOException
public void readExternal(ObjectInput in) throws IOException
IOException
public void writeExternal(ObjectOutput out) throws IOException
IOException
String emailId
Object value
long timestamp
int[] fields
TypeComparator<T> comparator
int keyLength
Object[] keyArray
TypeComparator<T> comparator
Object[] keyArray
boolean includeMetadata
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
DataInputDeserializer inputDeserializer
TypeInformation<T> fieldType
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
String msg
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
org.apache.flink.table.expressions.Expression alias
scala.collection.Seq<A> partitionBy
org.apache.flink.table.expressions.Expression orderBy
org.apache.flink.table.expressions.Expression preceding
org.apache.flink.table.expressions.Expression following
private Object readResolve()
private Object readResolve()
long minIdleStateRetentionTime
long maxIdleStateRetentionTime
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
String msg
private Object readResolve()
TypeSerializer<T> listSerializer
TypeInformation<T> elementType
TypeSerializer<T> mapSerializer
TypeInformation<T> keyType
TypeInformation<T> valueType
AggregateFunction<T,ACC> aggFunction
TypeInformation<T> evidence$1
TypeInformation<T> evidence$2
private Object readResolve()
org.apache.flink.table.expressions.Expression timeAttribute
String primaryKey
RowTypeInfo resultType
private Object readResolve()
TypeComparator<T> rowComp
TypeSerializer<T> rowSerializer
private Object readResolve()
RowTypeInfo rowType
private Object readResolve()
boolean cancel
boolean value
byte value
char value
double value
float value
int value
int fieldNumber
long value
int fieldPos
int fieldNumber
org.apache.flink.types.Record.InternalDeSerializer serializer
byte[] binaryData
byte[] switchBuffer
int[] offsets
int[] lengths
Value[] readFields
Value[] writeFields
int binaryLen
int numFields
int firstModifiedPos
Object[] fields
Serializable value
short value
char[] value
int len
int hashCode
long upperPart
long lowerPart
long to
long current
LongValue currentValue
long to
long current
private void readObject(ObjectInputStream stream) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream stream) throws IOException
IOException
Throwable failureCause
String id
TypeInformation<T> typeInfo
SerializableObject lock
int leaseCount
boolean closed
byte[] serializedData
String keyString
StringValue toTokenize
int pos
int limit
long seed
int status
int status
int exitCode
Copyright © 2014–2020 The Apache Software Foundation. All rights reserved.