boolean endReached
org.apache.hadoop.hbase.client.ResultScanner resultScanner
byte[] lastRow
int scannedRows
byte[] tableName
byte[] startRow
byte[] endRow
DataOutput out
ExecutionMode executionMode
boolean useClosureCleaner
int parallelism
int numberOfExecutionRetries
boolean forceKryo
boolean objectReuse
boolean autoTypeRegistrationEnabled
boolean forceAvro
CodeAnalysisMode codeAnalysisMode
boolean printProgressDuringExecution
long autoWatermarkInterval
long executionRetryDelay
RestartStrategies.RestartStrategyConfiguration restartStrategyConfiguration
long taskCancellationIntervalMillis
long taskCancellationTimeoutMillis
ExecutionConfig.GlobalJobParameters globalJobParameters
LinkedHashMap<K,V> registeredTypesWithKryoSerializers
LinkedHashMap<K,V> registeredTypesWithKryoSerializerClasses
LinkedHashMap<K,V> defaultKryoSerializers
LinkedHashMap<K,V> defaultKryoSerializerClasses
LinkedHashSet<E> registeredKryoTypes
LinkedHashSet<E> registeredPojoTypes
com.esotericsoftware.kryo.Serializer<T> serializer
long count
double sum
double localValue
double max
double min
int localValue
int max
int min
long localValue
long max
long min
DoubleValue wrapper
double sum
long sum
TypeComparator<T> typeComparator
Object[][] boundaries
TypeComparator<T>[] flatComparators
Object[] keys
long blockSize
long readRecords
long blockSize
long offset
byte[] delimiter
int lineLengthLimit
int bufferSize
int numLineSamples
Path filePath
long minSplitSize
int numSplits
long openTimeout
boolean unsplittable
boolean enumerateNestedFiles
Path outputFilePath
FileSystem.WriteMode writeMode
FileOutputFormat.OutputDirectoryMode outputDirectoryMode
boolean lineDelimiterIsLinebreak
Class<T>[] fieldTypes
boolean[] fieldIncluded
byte[] fieldDelim
boolean lenient
boolean skipFirstLineAsHeader
boolean quotedStringParsing
byte quoteCharacter
byte[] commentPrefix
int partitionNumber
InputFormat<OT,T extends InputSplit> replicatedIF
Map<K,V> fieldMapping1
Map<K,V> fieldMapping2
FieldSet readFields1
FieldSet readFields2
long count
BulkIterationBase.TerminationCriterionAggregator aggregator
Object userCodeObject
int restartAttempts
Time delayBetweenAttemptsInterval
FoldFunction<O,T> foldFunction
ReduceFunction<T> reduceFunction
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
String name
TypeSerializer<T> serializer
TimeUnit unit
long size
Class<T> arrayClass
TypeInformation<T> componentInfo
TypeComparator<T> comparator1
TypeComparator<T> comparator2
TypeComparator<T>[] comparators1
TypeComparator<T>[] comparators2
Object[] referenceKeyFields
Object[] candidateKeyFields
boolean ascendingComparison
TypeComparator<T>[] comparators
boolean ascendingComparison
BooleanValue reference
BooleanValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
ByteValue reference
ByteValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
CharValue reference
CharValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
DoubleValue reference
DoubleValue tempReference
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
boolean ascendingComparison
FloatValue reference
FloatValue tempReference
TypeComparator<T>[] comparators
Class<T> componentClass
TypeSerializer<T> componentSerializer
boolean ascendingComparison
IntValue reference
IntValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
LongValue reference
LongValue tempReference
TypeComparator<T>[] comparators
TypeComparator<T>[] comparators
boolean ascendingComparison
ShortValue reference
ShortValue tempReference
TypeComparator<T>[] comparators
boolean ascendingComparison
StringValue reference
StringValue tempReference
TypeComparator<T>[] comparators
TypeComparator<T>[] comparators
boolean ascending
BasicTypeComparator<T extends Comparable<T>> comparator
long count
long checksum
String id
long counter
long checksum
String id
TypeSerializer<T> serializer
SerializedListAccumulator<T> accumulator
String id
long counter
Comparable<T> value
Comparable<T> value
Comparable<T> value
Comparable<T> value
long agg
long agg
double agg
double agg
double agg
double agg
long agg
long agg
long agg
long agg
long agg
long agg
int count
TextOutputFormat.TextFormatter<IN> formatter
boolean withReplacement
int numSample
long seed
boolean withReplacement
int numSample
long seed
boolean withReplacement
double fraction
long seed
int[] fields
int[] fields
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.mapred.JobConf jobConf
org.apache.hadoop.mapred.OutputFormat<K,V> mapredOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.conf.Configuration configuration
org.apache.hadoop.mapreduce.OutputFormat<K,V> mapreduceOutputFormat
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TypeSerializer<T> serializer
TypeSerializer<T> typeSerializer
int id
SplittableIterator<T> source
String sinkIdentifier
boolean target
String partitionMarker
String charsetName
String charsetName
String charsetName
boolean skipInvalidLines
TupleSerializerBase<T> tupleSerializer
TypeSerializer<T> serializer
TypeSerializer<T> serializer
String username
String password
String drivername
String dbURL
String queryTemplate
int resultSetType
int resultSetConcurrency
RowTypeInfo rowTypeInfo
boolean hasNext
Object[][] parameterValues
String username
String password
String drivername
String dbURL
String query
int batchInterval
Connection dbConn
PreparedStatement upload
int batchCount
int[] typesArray
int[] fieldPositions
AggregationFunction<T>[] aggFunctions
int[] fields
boolean[] isFromFirst
Tuple outTuple
int[] fields
boolean[] isFromFirst
Tuple outTuple
CombineFunction<IN,OUT> wrappedFunction
TupleUnwrappingIterator<T,K> iter1
TupleUnwrappingIterator<T,K> iter2
TupleUnwrappingIterator<T,K> iter1
int[] fields
Tuple outTuple
TupleUnwrappingIterator<T,K> iter2
TupleUnwrappingIterator<T,K> iter
TupleUnwrappingIterator<T,K> iter
TupleWrappingCollector<IN,K> coll
TupleUnwrappingIterator<T,K> iter
Tuple3UnwrappingIterator<T,K1,K2> iter
Tuple3UnwrappingIterator<T,K1,K2> iter
Tuple3WrappingCollector<IN,K1,K2> coll
Tuple3UnwrappingIterator<T,K1,K2> iter
RichGroupReduceFunction<IN,OUT> wrappedFunction
KeySelector<IN,KEY> keySelector1
KeySelector<IN,KEY> keySelector2
Tuple3<T0,T1,T2> tuple
Function wrappedFunction
long trueCount
long falseCount
long nullCount
double value
double delta
double max
double min
CompensatedSum sum
float max
float min
CompensatedSum sum
int max
int min
int sum
long max
long min
long sum
long nonMissingCount
long nullCount
long nanCount
long infinityCount
Aggregator<T,R> min
Aggregator<T,R> max
Aggregator<T,R> sum
CompensatedSum mean
CompensatedSum m2
long nonNullCount
long nullCount
short max
short min
short sum
long nonNullCount
long nullCount
long emptyCount
int minStringLength
int maxStringLength
CompensatedSum meanLength
Aggregator<T,R>[] columnAggregators
Aggregator<T,R> aggregator
private Object readResolve() throws ObjectStreamException
ObjectStreamException
Object f0
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
Object f9
Object f10
Object f11
Object f12
Object f13
Object f14
Object f15
Object f16
Object f17
Object f18
Object f19
Object f20
Object f21
Object f22
Object f23
Object f24
Object f0
Object f1
Object f2
Object f3
Object f4
Object f5
Object f6
Object f7
Object f8
TypeInformation<T> leftType
TypeInformation<T> rightType
String functionName
InvalidTypesException typeException
Class<T> arrayType
TypeInformation<T> componentInfo
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeInformation<T> type
PojoField[] fields
int totalFields
String[] fieldNames
TypeInformation<T>[] types
int totalFields
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> leftSerializer
TypeSerializer<T> rightSerializer
boolean ascending
Class<T> type
TypeSerializer<T> serializer
TypeComparator<T>[] comparators
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T> serializer
Class<T> type
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
Class<T> clazz
TypeSerializer<T>[] fieldSerializers
int numFields
Map<K,V> registeredClasses
TypeSerializer<T>[] registeredSerializers
ExecutionConfig executionConfig
TypeComparator<T> comparator
TypeSerializer<T> serializer
boolean firstSerializer
Class<T> clazz
int[] keyPositions
TypeComparator<T>[] comparators
int[] normalizedKeyLengths
int numLeadingNormalizableKeys
int normalizableKeyPrefixLen
boolean invertNormKey
TypeSerializer<T>[] serializers
Class<T> tupleClass
TypeSerializer<T>[] fieldSerializers
int arity
int length
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
Class<T> type
boolean ascendingComparison
TypeComparator<T>[] comparators
LinkedHashMap<K,V> registeredTypesWithSerializers
LinkedHashMap<K,V> registeredTypesWithSerializerClasses
LinkedHashMap<K,V> defaultSerializers
LinkedHashMap<K,V> defaultSerializerClasses
LinkedHashSet<E> registeredTypes
Class<T> type
int[] fieldPositions
AggregationFunction<T>[] aggFunctions
TupleSerializerBase<T> serializer
TypeInformation<T> typeInfo
private void readObject(ObjectInputStream in)
String query
ClusterBuilder builder
String insertQuery
ClusterBuilder builder
private void readObject(ObjectInputStream ois) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
TypeSerializer<T> typeSerializer
int[] deweyNumber
private void readObject(ObjectInputStream ois) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
NonDuplicatingTypeSerializer<T> nonDuplicatingTypeSerializer
SharedBuffer<K extends Serializable,V> sharedBuffer
Set<E> states
long windowTime
boolean handleTimeout
int startEventCounter
private void readObject(ObjectInputStream ois) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
TypeSerializer<T> valueSerializer
String name
State.StateType stateType
Collection<E> stateTransitions
StateTransitionAction action
State<T> targetState
FilterFunction<T> condition
TypeSerializer<T> inputSerializer
boolean isProcessingTime
MultiplexingStreamRecordSerializer<T> streamRecordSerializer
NFA<T> nfa
KeySelector<IN,KEY> keySelector
TypeSerializer<T> keySerializer
org.apache.flink.cep.operator.AbstractKeyedCEPPatternOperator.PriorityQueueFactory<T> priorityQueueFactory
NFACompiler.NFAFactory<T> nfaFactory
FilterFunction<T> left
FilterFunction<T> right
Configuration backingConfig
String prefix
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
OperatorStatisticsConfig config
long cardinality
OperatorStatistics local
int countDbitmap
int countDlog2m
int heavyHitterSeed
double heavyHitterConfidence
double heavyHitterFraction
double heavyHitterError
boolean collectMin
boolean collectMax
boolean collectCountDistinct
boolean collectHeavyHitters
OperatorStatisticsConfig.CountDistinctAlgorithm countDistinctAlgorithm
OperatorStatisticsConfig.HeavyHitterAlgorithm heavyHitterAlgorithm
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream oos) throws IOException
IOException
Path checkpointDirectory
AbstractStateBackend nonPartitionedStateBackend
boolean fullyAsyncBackup
String operatorIdentifier
JobID jobId
Path[] configuredDbBasePaths
File[] initializedDbBasePaths
int nextDirectory
PredefinedOptions predefinedOptions
OptionsFactory optionsFactory
SerializableObject dbCleanupLock
Map<K,V> kvStateInformation
Path file
long start
long length
URI uri
int partitionNumber
int totalNumberOfPartitions
int splitNumber
String[] hostnames
int id
double x
double y
Collection<E> centroids
List<E> tasks
FileCopyTaskInputSplit curInputSplit
FileCopyTask task
int splitNumber
EnumTrianglesDataTypes.Edge outEdge
double dampening
double randomJump
double x
double y
double theta0
double theta1
Collection<E> parameters
LinearRegression.Params parameter
int count
EmptyFieldsCountAccumulator.VectorAccumulator emptyFieldCounter
Murmur3_32 hasher
long offset
double x
double y
long srcId
String message
long numberOfVertices
IterationRuntimeContext runtimeContext
long numberOfVertices
IterationRuntimeContext runtimeContext
double delta
TypeInformation<T> typeInformation
Object srcVertexId
TypeInformation<T> typeInformation
double beta
Object srcVertexId
Murmur3_32 hasher
Murmur3_32 hasher
Murmur3_32 hasher
Murmur3_32 hasher
Murmur3_32 hasher
Vertex<K,V> outVertex
Tuple2<T0,T1> outMsg
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
org.apache.flink.graph.pregel.ComputeFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
boolean setNewVertexValueCalled
Object first
MessageCombiner<K,Message> combinerFunction
long numberOfVertices
boolean optDegrees
IterationRuntimeContext runtimeContext
Collector<T> out
Collector<T> outWithDegrees
Vertex<K,V> outVal
Vertex<K,V> outValWithDegrees
long inDegree
long outDegree
boolean setNewVertexValueCalled
long numberOfVertices
EdgeDirection direction
Tuple2<T0,T1> outValue
IterationRuntimeContext runtimeContext
Iterator<E> edges
Collector<T> out
Object vertexId
org.apache.flink.graph.spargel.ScatterFunction.EdgesIterator<K,EV> edgeIterator
boolean edgesUsed
long inDegree
long outDegree
Object value
int seed
int count
int hash
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
org.apache.hadoop.conf.Configuration configuration
org.apache.hive.hcatalog.mapreduce.HCatInputFormat hCatInputFormat
org.apache.hadoop.mapreduce.RecordReader<KEYIN,VALUEIN> recordReader
boolean fetched
boolean hasNext
String[] fieldNames
org.apache.hive.hcatalog.data.schema.HCatSchema outputSchema
TypeInformation<T> resultType
scala.Function2<T1,T2,R> fun
Object broadcastVariable
scala.Function2<T1,T2,R> fun
Object broadcastVariable
scala.Function3<T1,T2,T3,R> fun
Object broadcastVariable
int fieldNumber
PythonStreamer streamer
PythonStreamer streamer
SerializationUtils.Serializer<IN> serializer
File inputFile
RandomAccessFile inputRAF
FileChannel inputChannel
MappedByteBuffer fileBuffer
boolean readAsByteArray
org.apache.flink.python.api.streaming.data.PythonReceiver.Deserializer<T> deserializer
File outputFile
RandomAccessFile outputRAF
FileChannel outputChannel
MappedByteBuffer fileBuffer
ByteBuffer[] saved
org.apache.flink.python.api.streaming.data.PythonSender.Serializer<T>[] serializer
int id
boolean usePython3
String planArguments
String inputFilePath
String outputFilePath
Process process
Thread shutdownThread
ServerSocket server
Socket socket
DataInputStream in
DataOutputStream out
int port
PythonSender<IN> sender
PythonReceiver receiver
StringBuilder msg
AbstractRichFunction function
DataInputStream input
DataOutputStream output
PythonPlanSender sender
PythonPlanReceiver receiver
Process process
ServerSocket server
Socket socket
JobID jobID
ExecutionAttemptID executionAttemptID
SerializedValue<T> flinkAccumulators
SerializedValue<T> userAccumulators
byte[] key
JobID job
long checkpointID
long timestamp
long duration
Map<K,V> taskStates
SerializedValue<T> keyGroupState
long stateSize
long duration
SerializedValue<T> state
long stateSize
long duration
JobVertexID jobVertexID
Map<K,V> subtaskStates
Map<K,V> kvStates
int parallelism
JobID jobID
long totalContainerMemoryMB
long taskManagerHeapSizeMB
long taskManagerDirectMemoryLimitMB
int numSlots
HashMap<K,V> taskManagerEnv
private Object readResolve() throws ObjectStreamException
ObjectStreamException
String message
SerializedThrowable error
private Object readResolve() throws ObjectStreamException
ObjectStreamException
int numRegisteredTaskManagers
int totalNumberOfSlots
ResourceID resourceID
akka.actor.ActorRef resourceManager
private Object readResolve() throws ObjectStreamException
ObjectStreamException
private Object readResolve() throws ObjectStreamException
ObjectStreamException
akka.actor.ActorRef resourceManager
akka.actor.ActorRef jobManager
Collection<E> currentlyRegisteredTaskManagers
ResourceID resourceId
ResourceID resourceId
String message
int numberOfWorkers
JobID jobId
ApplicationStatus finalStatus
String message
String jobManagerAddress
private Object readResolve() throws ObjectStreamException
ObjectStreamException
String resourceId
ResultPartitionID consumedPartitionId
ResultPartitionLocation consumedPartitionLocation
IntermediateDataSetID consumedResultId
int consumedSubpartitionIndex
DistributionPattern
and the subtask indices of the producing and consuming task.InputChannelDeploymentDescriptor[] inputChannels
IntermediateDataSetID resultId
IntermediateResultPartitionID partitionId
ResultPartitionType partitionType
int numberOfSubpartitions
boolean sendScheduleOrUpdateConsumersMessage
org.apache.flink.runtime.deployment.ResultPartitionLocation.LocationType locationType
ConnectionID connectionId
SerializedValue<T> serializedJobInformation
SerializedValue<T> serializedTaskInformation
ExecutionAttemptID executionId
int subtaskIndex
int attemptNumber
Collection<E> producedPartitions
Collection<E> inputGates
int targetSlotNumber
SerializedValue<T> operatorState
ExecutionVertex vertex
ExecutionAttemptID attemptId
long[] stateTimestamps
int attemptNumber
scala.concurrent.duration.FiniteDuration timeout
ConcurrentLinkedQueue<E> partialInputChannelDeploymentDescriptors
ExecutionState state
SimpleSlot assignedResource
Throwable failureCause
InstanceConnectionInfo assignedResourceLocation
SerializedValue<T> operatorState
Map<K,V> operatorKvState
scala.concurrent.ExecutionContext executionContext
SerializableObject accumulatorLock
Map<K,V> userAccumulators
Map<K,V> flinkAccumulators
SerializableObject progressLock
JobInformation jobInformation
SerializedValue<T> serializedJobInformation
boolean isStoppable
true
if all source tasks are stoppable.ConcurrentHashMap<K,V> tasks
List<E> verticesInCreationOrder
ConcurrentHashMap<K,V> intermediateResults
ConcurrentHashMap<K,V> currentExecutions
List<E> jobStatusListenerActors
List<E> executionListenerActors
long[] stateTimestamps
System.currentTimeMillis()
when
the execution graph transitioned into a certain state. The index into this array is the
ordinal of the enum value, i.e. the timestamp when the graph went into state "RUNNING" is
at stateTimestamps[RUNNING.ordinal()]
.scala.concurrent.duration.FiniteDuration timeout
boolean allowQueuedScheduling
ScheduleMode scheduleMode
boolean isArchived
JobStatus state
Throwable failureCause
int numFinishedJobVertices
Scheduler scheduler
RestartStrategy restartStrategy
ClassLoader userClassLoader
CheckpointCoordinator checkpointCoordinator
CheckpointStatsTracker checkpointStatsTracker
String jsonPlan
ExecutionConfigSummary executionConfigSummary
SerializableObject stateMonitor
ExecutionGraph graph
JobVertex jobVertex
ExecutionVertex[] taskVertices
IntermediateResult[] producedDataSets
List<E> inputs
int parallelism
boolean[] finishedSubtasks
int numSubtasksInFinalState
SlotSharingGroup slotSharingGroup
CoLocationGroup coLocationGroup
InputSplit[] inputSplits
List<E>[] inputSplitsPerSubtask
SerializedValue<T> serializedTaskInformation
InputSplitAssigner splitAssigner
ExecutionJobVertex jobVertex
Map<K,V> resultPartitions
ExecutionEdge[][] inputEdges
int subTaskIndex
EvictingBoundedList<T> priorExecutions
scala.concurrent.duration.FiniteDuration timeout
CoLocationConstraint locationConstraint
Execution currentExecution
List<E> locationConstraintInstances
boolean scheduleLocalOnly
JobID jobId
String jobName
SerializedValue<T> serializedExecutionConfig
Configuration jobConfiguration
Collection<E> requiredJarFileBlobKeys
Collection<E> requiredClasspathURLs
JobVertexID jobVertexId
String taskName
int parallelism
String invokableClassName
Configuration taskConfiguration
int maxAttempts
long delay
akka.actor.ActorRef actor
UUID leaderSessionID
MessageDecorator decorator
int numberOfCPUCores
long sizeOfPhysicalMemory
long sizeOfJvmHeap
long sizeOfManagedMemory
InetAddress inetAddress
int dataPort
String fqdnHostName
String hostName
Instance instance
boolean closed
InetSocketAddress address
int connectionIndex
SocketAddress address
ResultPartitionID partitionId
IntermediateResultPartitionID partitionId
ExecutionAttemptID producerId
int expectedSequenceNumber
int actualSequenceNumber
String formatDescription
IntermediateDataSetID id
JobVertex producer
List<E> consumers
ResultPartitionType resultType
boolean eagerlyDeployConsumers
If true
, the consumers are deployed as soon as the
runtime result is registered at the result manager of the task manager.
JobVertex target
DistributionPattern distributionPattern
IntermediateDataSet source
IntermediateDataSetID sourceId
String shipStrategyName
String preProcessingOperationName
String operatorLevelCachingDescription
Map<K,V> taskVertices
Configuration jobConfiguration
List<E> userJars
List<E> userJarBlobKeys
JobID jobID
String jobName
long sessionTimeout
boolean allowQueuedScheduling
ScheduleMode scheduleMode
JobSnapshottingSettings snapshotSettings
List<E> classpaths
SerializedValue<T> serializedExecutionConfig
SavepointRestoreSettings savepointRestoreSettings
JobVertexID id
ArrayList<E> results
ArrayList<E> inputs
int parallelism
Configuration configuration
String invokableClassName
boolean isStoppable
InputSplitSource<T extends InputSplit> inputSplitSource
String name
SlotSharingGroup slotSharingGroup
CoLocationGroup coLocationGroup
String operatorName
String operatorDescription
String operatorPrettyName
String resultOptimizerProperties
String formatDescription
String restorePath
boolean allowNonRestoredState
List<E> verticesToTrigger
List<E> verticesToAcknowledge
List<E> verticesToConfirm
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
AbstractID id
List<E> vertices
private Object readResolve() throws ObjectStreamException
ObjectStreamException
UUID leaderSessionID
JobID job
ExecutionAttemptID taskExecutionId
long checkpointId
SerializedValue<T> state
long stateSize
Throwable reason
long timestamp
long timestamp
int numJobsRunningOrPending
int numJobsFinished
int numJobsCancelled
int numJobsFailed
JobDetails[] runningJobs
JobDetails[] finishedJobs
boolean includeRunning
boolean includeFinished
private Object readResolve()
private Object readResolve()
private Object readResolve()
int numTaskManagersConnected
int numSlotsTotal
int numSlotsAvailable
TypeComparatorFactory<T> typeComparator
int parallelism
TypeComparatorFactory<T> comparatorFactory
Configuration config
TypeSerializer<T> elementSerializer
Serializable state
long checkpointId
byte[] serializedData
int numIds
Path filePath
TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
TypeSerializer<T> stateSerializer
StateDescriptor<S extends State,T> stateDesc
Path basePath
int fileStateThreshold
TypeSerializer<T> keySerializer
TypeSerializer<T> namespaceSerializer
TypeSerializer<T> stateSerializer
StateDescriptor<S extends State,T> stateDesc
byte[] data
byte[] data
int maxStateSize
byte[] serializedData
JobID jobID
ExecutionAttemptID executionId
ExecutionState executionState
SerializedThrowable throwable
AccumulatorSnapshot accumulators
String hostname
Configuration configuration
String[] tmpDirectories
byte[] buffer
int end
int position
backtype.storm.task.OutputCollector collector
String exclamation
boolean split
Random r
backtype.storm.spout.SpoutOutputCollector collector
boolean evenOrOdd
String token
backtype.storm.task.OutputCollector collector
StringBuilder lineBuilder
String prefix
OutputFormatter formatter
backtype.storm.spout.SpoutOutputCollector collector
String path
BufferedWriter writer
String path
BufferedReader reader
String line
boolean newLineRead
Object[] source
int counter
backtype.storm.topology.IRichSpout spout
org.apache.flink.storm.util.SpoutOutputCollectorObserver observer
backtype.storm.task.OutputCollector collector
backtype.storm.task.OutputCollector collector
String sentence
backtype.storm.topology.IRichBolt bolt
bolt
.String name
HashMap<K,V> numberOfAttributes
backtype.storm.generated.StormTopology stormTopology
HashMap<K,V> inputStreamIds
HashMap<K,V> inputComponentIds
HashMap<K,V> inputSchemas
HashMap<K,V> numberOfAttributes
backtype.storm.topology.IRichSpout spout
spout
.String name
boolean isRunning
Integer numberOfInvocations
ISpout.nextTuple()
calls.backtype.storm.generated.StormTopology stormTopology
CheckpointingMode checkpointingMode
long checkpointInterval
long checkpointTimeout
long minPauseBetweenCheckpoints
int maxConcurrentCheckpoints
boolean forceCheckpointing
long maxTimestamp
Comparator comparator
boolean byAggregate
boolean first
FieldAccessor<R,F> fieldAccessor
FieldAccessor<R,F> fieldAccessor
SumFunction adder
TypeSerializer<T> serializer
boolean isTuple
OutputFormat<IT> format
boolean cleanupCalled
boolean target
SerializableObject lock
SerializationSchema<T> schema
String hostName
int port
int maxNumRetries
boolean autoFlush
int retries
boolean isRunning
String path
ArrayList<E> tupleList
WriteFormat<IN> format
long millis
long lastTime
DeserializationSchema<T> schema
String path
int readerParallelism
FileInputFormat<OT> format
FileInputFormat
to be read.long interval
FileProcessingMode watchType
FileProcessingMode
.Long globalModificationTime
FilePathFilter pathFilter
boolean isRunning
FileInputFormat<OT> format
TypeSerializer<T> serializer
Tuple3<T0,T1,T2> readerState
TypeSerializer<T> serializer
byte[] elementsSerialized
int numElements
int numElementsEmitted
int numElementsToSkip
boolean isRunning
SplittableIterator<T> fullIterator
boolean isRunning
TypeInformation<T> typeInfo
InputFormat<OT,T extends InputSplit> format
boolean isRunning
TypeSerializer<T> idSerializer
long start
long end
long collected
boolean isRunning
long currentTimestamp
AscendingTimestampExtractor.MonotonyViolationHandler violationHandler
long currentMaxTimestamp
long lastEmittedWatermark
long maxOutOfOrderness
FoldFunction<O,T> foldFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
FoldFunction<O,T> foldFunction
byte[] serializedInitialValue
TypeSerializer<T> accSerializer
ReduceFunction<T> reduceFunction
AllWindowFunction<IN,OUT,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
WindowFunction<IN,OUT,KEY,W extends Window> windowFunction
ReduceFunction<T> reduceFunction
ReduceFunction<T> reduceFunction
int[] order
int fieldId
int fieldId
int[] indexes
Configuration config
String edgeId
StreamNode sourceVertex
StreamNode targetVertex
int typeNumber
List<E> selectedNames
StreamPartitioner<T> outputPartitioner
int id
Integer parallelism
Long bufferTimeout
String operatorName
String slotSharingGroup
KeySelector<IN,KEY> statePartitioner1
KeySelector<IN,KEY> statePartitioner2
TypeSerializer<T> stateKeySerializer
List<E> outputSelectors
TypeSerializer<T> typeSerializerIn1
TypeSerializer<T> typeSerializerIn2
TypeSerializer<T> typeSerializerOut
List<E> inEdges
List<E> outEdges
Class<T> jobVertexClass
InputFormat<OT,T extends InputSplit> inputFormat
String transformationId
ChainingStrategy chainingStrategy
AbstractStateBackend stateBackend
MetricGroup metrics
Function userFunction
Long count
byte[] serializedInitialValue
TypeSerializer<T> outTypeSerializer
TypeSerializer<T> serializer
TypeSerializer<T> outSerializer
int[] fields
int numFields
long combinedWatermark
long input1Watermark
long input2Watermark
long combinedWatermark
long input1Watermark
long input2Watermark
long sessionTimeout
long sessionTimeout
long size
long slide
long size
long slide
long size
long size
long maxCount
DeltaFunction<DATA> deltaFunction
double threshold
long windowSize
long interval
ReducingStateDescriptor<T> stateDesc
long interval
ReducingStateDescriptor<T> stateDesc
long maxCount
ReducingStateDescriptor<T> stateDesc
DeltaFunction<DATA> deltaFunction
double threshold
ValueStateDescriptor<T> stateDesc
ClusterBuilder builder
String keySpace
String table
long lastCommittedCheckpointID
ClusterBuilder builder
String insertQuery
String insertQuery
ClusterBuilder builder
Map<K,V> userConfig
List<E> transportNodes
IndexRequestBuilder<T> indexRequestBuilder
IndexRequest
from the incoming element.AtomicBoolean hasFailure
AtomicReference<V> failureThrowable
org.elasticsearch.action.bulk.BulkProcessor bulkProcessor
Map<K,V> userConfig
List<E> transportAddresses
ElasticsearchSinkFunction<T> elasticsearchSinkFunction
IndexRequest
from the incoming element.AtomicBoolean hasFailure
AtomicReference<V> failureThrowable
boolean initDone
String host
int port
SerializationSchema<T> schema
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
String formatString
long DEFAULT_BATCH_SIZE
String DEFAULT_IN_PROGRESS_SUFFIX
String DEFAULT_IN_PROGRESS_PREFIX
String DEFAULT_PENDING_SUFFIX
String DEFAULT_PENDING_PREFIX
String DEFAULT_VALID_SUFFIX
String DEFAULT_VALID_PREFIX
String DEFAULT_PART_REFIX
long DEFAULT_ASYNC_TIMEOUT_MS
String basePath
Path
that stored all rolling bucket directories.Bucketer bucketer
Bucketer
that is used to determine the path of bucket directories.Writer<T> writerTemplate
Writer<T> writer
long batchSize
boolean cleanupOnOpen
This should only be set to false if using the sink without checkpoints, to not remove the files already in the directory.
String inProgressSuffix
String inProgressPrefix
String pendingSuffix
String pendingPrefix
String validLengthSuffix
String validLengthPrefix
String partPrefix
long asyncTimeout
String charsetName
Properties kafkaProperties
long invalidOffsetBehavior
OffsetRequest
)long autoCommitInterval
Properties properties
long pollTimeout
KeyedDeserializationSchema<T> deserializer
List<E> allSubscribedPartitions
SerializedValue<T> periodicWatermarkAssigner
SerializedValue<T> punctuatedWatermarkAssigner
org.apache.commons.collections.map.LinkedMap pendingCheckpoints
boolean running
int[] partitions
Properties producerConfig
String defaultTopicId
KeyedSerializationSchema<T> schema
KafkaPartitioner<T> partitioner
boolean logFailuresOnly
boolean flushOnCheckpoint
SerializableObject pendingRecordsLock
long pendingRecords
String topic
int partition
int cachedHash
int leaderId
int leaderPort
String leaderHost
KafkaTopicPartition topicPartition
int cachedHash
int targetPartition
org.apache.nifi.remote.client.SiteToSiteClient client
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
NiFiDataPacketBuilder<T> builder
org.apache.nifi.remote.client.SiteToSiteClientConfig clientConfig
long waitTimeMs
boolean isRunning
String queueName
RMQConnectionConfig rmqConnectionConfig
SerializationSchema<T> schema
boolean logFailuresOnly
RMQConnectionConfig rmqConnectionConfig
String queueName
boolean usesCorrelationId
DeserializationSchema<T> schema
String host
Integer port
String virtualHost
String username
String password
String uri
Integer networkRecoveryInterval
Boolean automaticRecovery
Boolean topologyRecovery
Integer connectionTimeout
Integer requestedChannelMax
Integer requestedFrameMax
Integer requestedHeartbeat
String additionalKey
RedisDataType.HASH
and RedisDataType.SORTED_SET
.
Other RedisDataType
works only with two variable i.e. name of the list and value to be added.
But for RedisDataType.HASH
and RedisDataType.SORTED_SET
we need three variables.
For RedisDataType.HASH
we need hash name, hash key and element.
additionalKey
used as hash name for RedisDataType.HASH
For RedisDataType.SORTED_SET
we need set name, the element and it's score.
additionalKey
used as set name for RedisDataType.SORTED_SET
RedisMapper<T> redisSinkMapper
RedisCommand redisCommand
FlinkJedisConfigBase flinkJedisConfigBase
RedisCommandsContainer redisCommandsContainer
int maxTotal
int maxIdle
int minIdle
int connectionTimeout
redis.clients.jedis.JedisCluster jedisCluster
redis.clients.jedis.JedisPool jedisPool
redis.clients.jedis.JedisSentinelPool jedisSentinelPool
RedisCommand redisCommand
String additionalKey
RedisDataType.HASH
and RedisDataType.SORTED_SET
.
Other RedisDataType
works only with two variable i.e. name of the list and value to be added.
But for RedisDataType.HASH
and RedisDataType.SORTED_SET
we need three variables.
For RedisDataType.HASH
we need hash name, hash key and element.
RedisCommandDescription.getAdditionalKey()
used as hash name for RedisDataType.HASH
For RedisDataType.SORTED_SET
we need set name, the element and it's score.
RedisCommandDescription.getAdditionalKey()
used as set name for RedisDataType.SORTED_SET
Properties properties
TwitterSource.EndpointInitializer initializer
Random rnd
Random rnd
int counter
int counter
long counter
CheckpointCommitter committer
TypeSerializer<T> serializer
String id
GenericWriteAheadSink.ExactlyOnceState state
long currentWatermark
Function function
KeySelector<IN,KEY> keySelector
TypeSerializer<T> keySerializer
TypeSerializer<T> stateTypeSerializer
long windowSize
long windowSlide
long paneSize
int numPanesPerWindow
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
WindowAssigner<T,W extends Window> windowAssigner
KeySelector<IN,KEY> keySelector
Trigger<T,W extends Window> trigger
StateDescriptor<S extends State,T> windowStateDescriptor
TypeSerializer<T> inputSerializer
TypeSerializer<T> keySerializer
TypeSerializer<T> windowSerializer
long allowedLateness
window.maxTimestamp + allowedLateness
landmark.
AllWindowFunction<IN,OUT,W extends Window> wrappedFunction
AllWindowFunction<IN,OUT,W extends Window> wrappedFunction
int[] returnArray
boolean set
int setNumber
int[] returnArray
Partitioner<K> partitioner
KeySelector<IN,KEY> keySelector
int[] returnArray
int[] returnArray
int[] returnArray
KeySelector<IN,KEY> keySelector
int[] returnArray
int[] returnArray
Random random
int[] returnArray
TypeSerializer<T> typeSerializer
TypeSerializer<T> typeSerializer
StateHandle<T> operatorState
StateHandle<T> functionState
HashMap<K,V> kvStates
StreamTaskState[] states
TypeInformation<T> fieldType
int pos
PojoComparator<T> comparator
int pos
TupleSerializerBase<T> serializer
Object[] fields
int length
int pos
TypeComparator<T> comparator
int keyLength
Object[] keyArray
TypeComparator<T> comparator
Object[] keyArray
com.fasterxml.jackson.databind.ObjectMapper mapper
boolean includeMetadata
com.fasterxml.jackson.databind.ObjectMapper mapper
String[] fieldNames
TypeInformation<T>[] fieldTypes
com.fasterxml.jackson.databind.ObjectMapper objectMapper
boolean failOnMissingField
DeserializationSchema<T> deserializationSchema
SerializationSchema<T> serializationSchema
TypeSerializer<T> keySerializer
TypeSerializer<T> valueSerializer
DataInputDeserializer inputDeserializer
TypeSerializer<T> serializer
boolean value
byte value
char value
double value
float value
int value
int fieldNumber
long value
int fieldPos
int fieldNumber
org.apache.flink.types.Record.InternalDeSerializer serializer
byte[] binaryData
byte[] switchBuffer
int[] offsets
int[] lengths
Value[] readFields
Value[] writeFields
int binaryLen
int numFields
int firstModifiedPos
short value
char[] value
int len
int hashCode
long upperPart
long lowerPart
String toString
long to
long current
LongValue currentValue
long to
long current
byte[] serializedData
StringValue toTokenize
int pos
int limit
long seed
Copyright © 2014–2017 The Apache Software Foundation. All rights reserved.