Package | Description |
---|---|
org.apache.flink.connectors.hive | |
org.apache.flink.connectors.hive.read | |
org.apache.flink.connectors.hive.util |
Modifier and Type | Method and Description |
---|---|
HiveTablePartition |
HiveTablePartitionSerializer.deserialize(int version,
byte[] serialized) |
HiveTablePartition |
HiveTablePartitionSplits.getHiveTablePartition() |
static HiveTablePartition |
HiveTablePartition.ofPartition(HiveConf hiveConf,
String hiveVersion,
String dbName,
String tableName,
LinkedHashMap<String,String> partitionSpec)
Creates a HiveTablePartition to represent a hive partition.
|
static HiveTablePartition |
HiveTablePartition.ofTable(HiveConf hiveConf,
String hiveVersion,
String dbName,
String tableName)
Creates a HiveTablePartition to represent a hive table.
|
HiveTablePartition |
HiveTableSource.HiveContinuousPartitionFetcherContext.toHiveTablePartition(org.apache.hadoop.hive.metastore.api.Partition partition)
Convert partition to HiveTablePartition.
|
Modifier and Type | Method and Description |
---|---|
byte[] |
HiveTablePartitionSerializer.serialize(HiveTablePartition hiveTablePartition) |
Modifier and Type | Method and Description |
---|---|
static List<HiveSourceSplit> |
HiveSourceFileEnumerator.createInputSplits(int minNumSplits,
List<HiveTablePartition> partitions,
org.apache.hadoop.mapred.JobConf jobConf,
boolean isForParallelismInfer) |
List<HiveTablePartitionSplits> |
MRSplitsGetter.getHiveTablePartitionMRSplits(int minNumSplits,
List<HiveTablePartition> partitions,
org.apache.hadoop.mapred.JobConf jobConf) |
static int |
HiveSourceFileEnumerator.getNumFiles(List<HiveTablePartition> partitions,
org.apache.hadoop.mapred.JobConf jobConf) |
HiveSourceBuilder |
HiveSourceBuilder.setPartitions(List<HiveTablePartition> partitions)
Sets the partitions to read in batch mode.
|
Constructor and Description |
---|
HiveTablePartitionSplits(HiveTablePartition hiveTablePartition,
org.apache.hadoop.mapred.JobConf jobConf,
org.apache.hadoop.mapred.InputSplit[] inputSplits) |
PartitionFilesSizeCalculator(HiveTablePartition hiveTablePartition,
Long openCost,
org.apache.hadoop.mapred.JobConf jobConf) |
Constructor and Description |
---|
HiveSourceDynamicFileEnumerator(String table,
List<String> dynamicFilterPartitionKeys,
List<HiveTablePartition> allPartitions,
String hiveVersion,
org.apache.hadoop.mapred.JobConf jobConf) |
HiveSourceFileEnumerator(List<HiveTablePartition> partitions,
org.apache.hadoop.mapred.JobConf jobConf) |
Modifier and Type | Field and Description |
---|---|
protected HiveTablePartition |
HiveTableInputSplit.hiveTablePartition |
Modifier and Type | Method and Description |
---|---|
HiveTablePartition |
HiveTableInputSplit.getHiveTablePartition() |
HiveTablePartition |
HiveSourceSplit.getHiveTablePartition() |
HiveTablePartition |
HivePartitionFetcherContextBase.toHiveTablePartition(P partition) |
HiveTablePartition |
HivePartitionContext.toHiveTablePartition(P partition)
Convert partition to
HiveTablePartition . |
Modifier and Type | Method and Description |
---|---|
static HiveTableInputSplit[] |
HiveTableInputFormat.createInputSplits(int minNumSplits,
List<HiveTablePartition> partitions,
int threadNum,
org.apache.hadoop.mapred.JobConf jobConf) |
void |
HiveInputFormatPartitionReader.open(List<HiveTablePartition> partitions) |
Constructor and Description |
---|
HiveSourceSplit(org.apache.hadoop.mapred.FileSplit fileSplit,
HiveTablePartition hiveTablePartition,
CheckpointedPosition readerPosition) |
HiveSourceSplit(String id,
Path filePath,
long offset,
long length,
long fileModificationTime,
long fileSize,
String[] hostnames,
CheckpointedPosition readerPosition,
HiveTablePartition hiveTablePartition) |
HiveSourceSplit(String id,
Path filePath,
long offset,
long length,
String[] hostnames,
CheckpointedPosition readerPosition,
HiveTablePartition hiveTablePartition)
Deprecated.
|
HiveTableInputSplit(int splitNumber,
org.apache.hadoop.mapred.InputSplit hInputSplit,
org.apache.hadoop.mapred.JobConf jobconf,
HiveTablePartition hiveTablePartition) |
Constructor and Description |
---|
HiveTableInputFormat(int threadNum,
org.apache.hadoop.mapred.JobConf jobConf,
List<String> partitionKeys,
DataType[] fieldTypes,
String[] fieldNames,
int[] projectedFields,
Long limit,
String hiveVersion,
boolean useMapRedReader,
List<HiveTablePartition> partitions) |
Modifier and Type | Method and Description |
---|---|
static HiveTablePartition |
HivePartitionUtils.toHiveTablePartition(List<String> partitionKeys,
Properties tableProps,
org.apache.hadoop.hive.metastore.api.Partition partition) |
Modifier and Type | Method and Description |
---|---|
static List<HiveTablePartition> |
HivePartitionUtils.deserializeHiveTablePartition(List<byte[]> partitionBytes) |
static List<HiveTablePartition> |
HivePartitionUtils.getAllPartitions(org.apache.hadoop.mapred.JobConf jobConf,
String hiveVersion,
ObjectPath tablePath,
List<String> partitionColNames,
List<Map<String,String>> remainingPartitions)
Returns all HiveTablePartitions of a hive table, returns single HiveTablePartition if the
hive table is not partitioned.
|
Modifier and Type | Method and Description |
---|---|
static List<byte[]> |
HivePartitionUtils.serializeHiveTablePartition(List<HiveTablePartition> hiveTablePartitions) |
Copyright © 2014–2024 The Apache Software Foundation. All rights reserved.