public class SparkTable extends Object implements org.apache.spark.sql.connector.catalog.Table, org.apache.spark.sql.connector.catalog.SupportsRead, org.apache.spark.sql.connector.catalog.SupportsWrite, org.apache.spark.sql.connector.catalog.SupportsDelete
Table
for table store.Constructor and Description |
---|
SparkTable(Table table,
Lock.Factory lockFactory,
org.apache.flink.configuration.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
Set<org.apache.spark.sql.connector.catalog.TableCapability> |
capabilities() |
void |
deleteWhere(org.apache.spark.sql.sources.Filter[] filters) |
String |
name() |
org.apache.spark.sql.connector.read.ScanBuilder |
newScanBuilder(org.apache.spark.sql.util.CaseInsensitiveStringMap options) |
org.apache.spark.sql.connector.write.WriteBuilder |
newWriteBuilder(org.apache.spark.sql.connector.write.LogicalWriteInfo info) |
org.apache.spark.sql.connector.expressions.Transform[] |
partitioning() |
Map<String,String> |
properties() |
org.apache.spark.sql.types.StructType |
schema() |
public SparkTable(Table table, Lock.Factory lockFactory, org.apache.flink.configuration.Configuration conf)
public org.apache.spark.sql.connector.read.ScanBuilder newScanBuilder(org.apache.spark.sql.util.CaseInsensitiveStringMap options)
newScanBuilder
in interface org.apache.spark.sql.connector.catalog.SupportsRead
public String name()
name
in interface org.apache.spark.sql.connector.catalog.Table
public org.apache.spark.sql.types.StructType schema()
schema
in interface org.apache.spark.sql.connector.catalog.Table
public Set<org.apache.spark.sql.connector.catalog.TableCapability> capabilities()
capabilities
in interface org.apache.spark.sql.connector.catalog.Table
public org.apache.spark.sql.connector.expressions.Transform[] partitioning()
partitioning
in interface org.apache.spark.sql.connector.catalog.Table
public org.apache.spark.sql.connector.write.WriteBuilder newWriteBuilder(org.apache.spark.sql.connector.write.LogicalWriteInfo info)
newWriteBuilder
in interface org.apache.spark.sql.connector.catalog.SupportsWrite
public void deleteWhere(org.apache.spark.sql.sources.Filter[] filters)
deleteWhere
in interface org.apache.spark.sql.connector.catalog.SupportsDelete
Copyright © 2019–2023 The Apache Software Foundation. All rights reserved.