Package | Description |
---|---|
org.apache.hadoop.hive.accumulo |
Serde and InputFormat support for connecting Hive to Accumulo tables.
|
org.apache.hadoop.hive.hbase |
Implements an HBase storage handler for Hive.
|
org.apache.hadoop.hive.metastore | |
org.apache.hadoop.hive.metastore.api | |
org.apache.hadoop.hive.metastore.events | |
org.apache.hadoop.hive.metastore.parser |
This package implements the parser for parsing the string filter
for the listPartitionsByFilter API.
|
org.apache.hadoop.hive.ql.hooks | |
org.apache.hadoop.hive.ql.index | |
org.apache.hadoop.hive.ql.index.bitmap | |
org.apache.hadoop.hive.ql.index.compact | |
org.apache.hadoop.hive.ql.metadata | |
org.apache.hadoop.hive.ql.parse | |
org.apache.hadoop.hive.ql.security.authorization | |
org.apache.hive.hcatalog.messaging | |
org.apache.hive.hcatalog.messaging.json | |
org.apache.hive.hcatalog.streaming |
HCatalog Streaming API -- high level description
|
Modifier and Type | Method and Description |
---|---|
void |
AccumuloStorageHandler.commitCreateTable(Table table) |
void |
AccumuloStorageHandler.commitDropTable(Table table,
boolean deleteData) |
protected String |
AccumuloStorageHandler.getTableName(Table table) |
protected boolean |
AccumuloStorageHandler.isExternalTable(Table table) |
void |
AccumuloStorageHandler.preCreateTable(Table table) |
void |
AccumuloStorageHandler.preDropTable(Table table) |
void |
AccumuloStorageHandler.rollbackCreateTable(Table table) |
void |
AccumuloStorageHandler.rollbackDropTable(Table table) |
Modifier and Type | Method and Description |
---|---|
void |
HBaseStorageHandler.commitCreateTable(Table table) |
void |
HBaseStorageHandler.commitDropTable(Table tbl,
boolean deleteData) |
void |
HBaseStorageHandler.preCreateTable(Table tbl) |
void |
HBaseStorageHandler.preDropTable(Table table) |
void |
HBaseStorageHandler.rollbackCreateTable(Table table) |
void |
HBaseStorageHandler.rollbackDropTable(Table table) |
Modifier and Type | Method and Description |
---|---|
static Table |
MetaStoreUtils.createColumnsetSchema(String name,
List<String> columns,
List<String> partCols,
org.apache.hadoop.conf.Configuration conf) |
protected Table |
HiveMetaStoreClient.deepCopy(Table table) |
Table |
MetaStoreFilterHook.filterTable(Table table)
filter to given table object if applicable
|
Table |
DefaultMetaStoreFilterHookImpl.filterTable(Table table) |
Table |
HiveMetaStore.HMSHandler.get_table_core(String dbname,
String name)
Equivalent of get_table, but does not log audits and fire pre-event listener.
|
Table |
HiveMetaStore.HMSHandler.get_table(String dbname,
String name) |
Table |
IMetaStoreClient.getTable(String tableName)
Deprecated.
As of release 0.6.0 replaced by
IMetaStoreClient.getTable(String, String) .
This method will be removed in release 0.7.0. |
Table |
HiveMetaStoreClient.getTable(String tableName)
Deprecated.
|
Table |
RawStore.getTable(String dbName,
String tableName) |
Table |
ObjectStore.getTable(String dbName,
String tableName) |
Table |
IMetaStoreClient.getTable(String dbName,
String tableName)
Get a table object.
|
Table |
HiveMetaStoreClient.getTable(String dbname,
String name) |
Table |
RawStore.markPartitionForEvent(String dbName,
String tblName,
Map<String,String> partVals,
PartitionEventType evtType) |
Table |
ObjectStore.markPartitionForEvent(String dbName,
String tblName,
Map<String,String> partName,
PartitionEventType evtType) |
Modifier and Type | Method and Description |
---|---|
List<Table> |
MetaStoreFilterHook.filterTables(List<Table> tableList)
Filter given list of tables
|
List<Table> |
DefaultMetaStoreFilterHookImpl.filterTables(List<Table> tableList) |
List<Table> |
HiveMetaStore.HMSHandler.get_table_objects_by_name(String dbname,
List<String> names)
Gets multiple tables from the hive metastore.
|
List<Table> |
RawStore.getTableObjectsByName(String dbname,
List<String> tableNames) |
List<Table> |
ObjectStore.getTableObjectsByName(String db,
List<String> tbl_names) |
List<Table> |
IMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
List<Table> |
HiveMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
Modifier and Type | Method and Description |
---|---|
Index |
HiveMetaStore.HMSHandler.add_index(Index newIndex,
Table indexTable) |
void |
HiveMetaStore.HMSHandler.alter_table_with_cascade(String dbname,
String name,
Table newTable,
boolean cascade) |
void |
HiveMetaStore.HMSHandler.alter_table_with_environment_context(String dbname,
String name,
Table newTable,
EnvironmentContext envContext) |
void |
IMetaStoreClient.alter_table(String defaultDatabaseName,
String tblName,
Table table) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
HiveMetaStore.HMSHandler.alter_table(String dbname,
String name,
Table newTable) |
void |
IMetaStoreClient.alter_table(String defaultDatabaseName,
String tblName,
Table table,
boolean cascade) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext envContext) |
void |
HiveAlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newt) |
void |
AlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newTable)
handles alter table
|
void |
HiveAlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newt,
boolean cascade) |
void |
AlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newTable,
boolean cascade)
handles alter table, the changes could be cascaded to partitions if applicable
|
void |
RawStore.alterTable(String dbname,
String name,
Table newTable) |
void |
ObjectStore.alterTable(String dbname,
String name,
Table newTable) |
static boolean |
MetaStoreUtils.canDropPartition(Table table,
Partition partition) |
void |
HiveMetaHook.commitCreateTable(Table table)
Called after successfully adding a new table definition to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.commitDropTable(Table table,
boolean deleteData)
Called after successfully removing a table definition from the metastore
during DROP TABLE.
|
protected void |
HiveMetaStoreClient.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.create_table(Table tbl) |
void |
IMetaStoreClient.createIndex(Index index,
Table indexTable)
create an index
|
void |
HiveMetaStoreClient.createIndex(Index index,
Table indexTable)
create an index
|
void |
RawStore.createTable(Table tbl) |
void |
ObjectStore.createTable(Table tbl) |
void |
IMetaStoreClient.createTable(Table tbl) |
void |
HiveMetaStoreClient.createTable(Table tbl) |
void |
HiveMetaStoreClient.createTable(Table tbl,
EnvironmentContext envContext) |
protected Table |
HiveMetaStoreClient.deepCopy(Table table) |
Table |
MetaStoreFilterHook.filterTable(Table table)
filter to given table object if applicable
|
Table |
DefaultMetaStoreFilterHookImpl.filterTable(Table table) |
static Deserializer |
MetaStoreUtils.getDeserializer(org.apache.hadoop.conf.Configuration conf,
Partition part,
Table table)
getDeserializer
Get the Deserializer for a partition.
|
static Deserializer |
MetaStoreUtils.getDeserializer(org.apache.hadoop.conf.Configuration conf,
Table table,
boolean skipConfError)
getDeserializer
Get the Deserializer for a table.
|
static Class<? extends Deserializer> |
MetaStoreUtils.getDeserializerClass(org.apache.hadoop.conf.Configuration conf,
Table table) |
org.apache.hadoop.fs.FileStatus[] |
Warehouse.getFileStatusesForUnpartitionedTable(Database db,
Table table) |
HiveMetaHook |
HiveMetaHookLoader.getHook(Table tbl)
Loads a hook for the specified table.
|
static Properties |
MetaStoreUtils.getPartitionMetadata(Partition partition,
Table table) |
static ProtectMode |
MetaStoreUtils.getProtectMode(Table table) |
static String |
Warehouse.getQualifiedName(Table table) |
static Properties |
MetaStoreUtils.getSchema(Partition part,
Table table) |
static Properties |
MetaStoreUtils.getTableMetadata(Table table) |
static boolean |
MetaStoreUtils.isExternalTable(Table table)
Determines whether a table is an external table.
|
static boolean |
MetaStoreUtils.isImmutableTable(Table table)
Determines whether a table is an immutable table.
|
static boolean |
MetaStoreUtils.isIndexTable(Table table) |
static boolean |
MetaStoreUtils.isNonNativeTable(Table table) |
static boolean |
MetaStoreUtils.isView(Table table) |
void |
HiveMetaHook.preCreateTable(Table table)
Called before a new table definition is added to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.preDropTable(Table table)
Called before a table definition is removed from the metastore
during DROP TABLE.
|
static boolean |
MetaStoreUtils.requireCalStats(org.apache.hadoop.conf.Configuration hiveConf,
Partition oldPart,
Partition newPart,
Table tbl) |
void |
HiveMetaHook.rollbackCreateTable(Table table)
Called after failure adding a new table definition to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.rollbackDropTable(Table table)
Called after failure removing a table definition from the metastore
during DROP TABLE.
|
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Database db,
Table tbl,
Warehouse wh,
boolean madeDir) |
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Database db,
Table tbl,
Warehouse wh,
boolean madeDir,
boolean forceRecompute) |
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Table tbl,
org.apache.hadoop.fs.FileStatus[] fileStatus,
boolean newDir,
boolean forceRecompute)
Updates the numFiles and totalSize parameters for the passed unpartitioned Table by querying
the warehouse if the passed Table does not already have values for these parameters.
|
Modifier and Type | Method and Description |
---|---|
List<Table> |
MetaStoreFilterHook.filterTables(List<Table> tableList)
Filter given list of tables
|
List<Table> |
DefaultMetaStoreFilterHookImpl.filterTables(List<Table> tableList) |
Modifier and Type | Method and Description |
---|---|
Table |
Table.deepCopy() |
Table |
ThriftHiveMetastore.Iface.get_table(String dbname,
String tbl_name) |
Table |
ThriftHiveMetastore.Client.get_table(String dbname,
String tbl_name) |
Table |
ThriftHiveMetastore.add_index_args.getIndex_table() |
Table |
ThriftHiveMetastore.alter_table_args.getNew_tbl() |
Table |
ThriftHiveMetastore.alter_table_with_environment_context_args.getNew_tbl() |
Table |
ThriftHiveMetastore.alter_table_with_cascade_args.getNew_tbl() |
Table |
ThriftHiveMetastore.AsyncClient.get_table_call.getResult() |
Table |
ThriftHiveMetastore.get_table_result.getSuccess() |
Table |
ThriftHiveMetastore.create_table_args.getTbl() |
Table |
ThriftHiveMetastore.create_table_with_environment_context_args.getTbl() |
Table |
ThriftHiveMetastore.Client.recv_get_table() |
Modifier and Type | Method and Description |
---|---|
List<Table> |
ThriftHiveMetastore.Iface.get_table_objects_by_name(String dbname,
List<String> tbl_names) |
List<Table> |
ThriftHiveMetastore.Client.get_table_objects_by_name(String dbname,
List<String> tbl_names) |
List<Table> |
ThriftHiveMetastore.AsyncClient.get_table_objects_by_name_call.getResult() |
List<Table> |
ThriftHiveMetastore.get_table_objects_by_name_result.getSuccess() |
Iterator<Table> |
ThriftHiveMetastore.get_table_objects_by_name_result.getSuccessIterator() |
List<Table> |
ThriftHiveMetastore.Client.recv_get_table_objects_by_name() |
Modifier and Type | Method and Description |
---|---|
Index |
ThriftHiveMetastore.Iface.add_index(Index new_index,
Table index_table) |
Index |
ThriftHiveMetastore.Client.add_index(Index new_index,
Table index_table) |
void |
ThriftHiveMetastore.AsyncIface.add_index(Index new_index,
Table index_table,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.add_index_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.add_index(Index new_index,
Table index_table,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.add_index_call> resultHandler) |
void |
ThriftHiveMetastore.get_table_objects_by_name_result.addToSuccess(Table elem) |
void |
ThriftHiveMetastore.Iface.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
ThriftHiveMetastore.Client.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
ThriftHiveMetastore.AsyncIface.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_cascade_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_cascade_call> resultHandler) |
void |
ThriftHiveMetastore.Iface.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.AsyncIface.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context_call> resultHandler) |
void |
ThriftHiveMetastore.Iface.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
ThriftHiveMetastore.Client.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
ThriftHiveMetastore.AsyncIface.alter_table(String dbname,
String tbl_name,
Table new_tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_call> resultHandler) |
int |
Table.compareTo(Table other) |
void |
ThriftHiveMetastore.Iface.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.AsyncIface.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_with_environment_context_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_with_environment_context_call> resultHandler) |
void |
ThriftHiveMetastore.Iface.create_table(Table tbl) |
void |
ThriftHiveMetastore.Client.create_table(Table tbl) |
void |
ThriftHiveMetastore.AsyncIface.create_table(Table tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_call> resultHandler) |
void |
ThriftHiveMetastore.AsyncClient.create_table(Table tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_call> resultHandler) |
boolean |
Table.equals(Table that) |
void |
ThriftHiveMetastore.Client.send_add_index(Index new_index,
Table index_table) |
void |
ThriftHiveMetastore.Client.send_alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
ThriftHiveMetastore.Client.send_alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.send_alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
ThriftHiveMetastore.Client.send_create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.send_create_table(Table tbl) |
void |
ThriftHiveMetastore.add_index_args.setIndex_table(Table index_table) |
void |
ThriftHiveMetastore.alter_table_args.setNew_tbl(Table new_tbl) |
void |
ThriftHiveMetastore.alter_table_with_environment_context_args.setNew_tbl(Table new_tbl) |
void |
ThriftHiveMetastore.alter_table_with_cascade_args.setNew_tbl(Table new_tbl) |
void |
ThriftHiveMetastore.get_table_result.setSuccess(Table success) |
void |
ThriftHiveMetastore.create_table_args.setTbl(Table tbl) |
void |
ThriftHiveMetastore.create_table_with_environment_context_args.setTbl(Table tbl) |
Modifier and Type | Method and Description |
---|---|
void |
ThriftHiveMetastore.get_table_objects_by_name_result.setSuccess(List<Table> success) |
Constructor and Description |
---|
Table(Table other)
Performs a deep copy on other.
|
ThriftHiveMetastore.add_index_args(Index new_index,
Table index_table) |
ThriftHiveMetastore.alter_table_args(String dbname,
String tbl_name,
Table new_tbl) |
ThriftHiveMetastore.alter_table_with_cascade_args(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
ThriftHiveMetastore.alter_table_with_environment_context_args(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
ThriftHiveMetastore.AsyncClient.add_index_call(Index new_index,
Table index_table,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.add_index_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.AsyncClient.alter_table_call(String dbname,
String tbl_name,
Table new_tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.AsyncClient.alter_table_with_cascade_call(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_cascade_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context_call(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.AsyncClient.create_table_call(Table tbl,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.AsyncClient.create_table_with_environment_context_call(Table tbl,
EnvironmentContext environment_context,
org.apache.thrift.async.AsyncMethodCallback<ThriftHiveMetastore.AsyncClient.create_table_with_environment_context_call> resultHandler,
org.apache.thrift.async.TAsyncClient client,
org.apache.thrift.protocol.TProtocolFactory protocolFactory,
org.apache.thrift.transport.TNonblockingTransport transport) |
ThriftHiveMetastore.create_table_args(Table tbl) |
ThriftHiveMetastore.create_table_with_environment_context_args(Table tbl,
EnvironmentContext environment_context) |
ThriftHiveMetastore.get_table_result(Table success,
MetaException o1,
NoSuchObjectException o2) |
Constructor and Description |
---|
ThriftHiveMetastore.get_table_objects_by_name_result(List<Table> success,
MetaException o1,
InvalidOperationException o2,
UnknownDBException o3) |
Modifier and Type | Method and Description |
---|---|
Table |
PreAlterTableEvent.getNewTable() |
Table |
AlterTableEvent.getNewTable() |
Table |
PreAlterTableEvent.getOldTable() |
Table |
AlterTableEvent.getOldTable() |
Table |
PreReadTableEvent.getTable() |
Table |
PreDropTableEvent.getTable() |
Table |
PreDropPartitionEvent.getTable() |
Table |
PreCreateTableEvent.getTable() |
Table |
PreAddPartitionEvent.getTable() |
Table |
LoadPartitionDoneEvent.getTable() |
Table |
DropTableEvent.getTable() |
Table |
DropPartitionEvent.getTable() |
Table |
CreateTableEvent.getTable() |
Table |
AlterPartitionEvent.getTable()
Get the table this partition is in
|
Table |
AddPartitionEvent.getTable() |
Modifier and Type | Method and Description |
---|---|
void |
ExpressionTree.TreeNode.generateJDOFilter(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuffer)
Generates a JDO filter statement
|
void |
ExpressionTree.LeafNode.generateJDOFilter(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuilder) |
void |
ExpressionTree.generateJDOFilterFragment(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuilder)
Generate the JDOQL filter for the given expression tree
|
int |
ExpressionTree.LeafNode.getPartColIndexForFilter(Table table,
ExpressionTree.FilterBuilder filterBuilder)
Get partition column index in the table partition column list that
corresponds to the key that is being filtered on by this tree node.
|
Modifier and Type | Method and Description |
---|---|
Table |
LineageInfo.DataContainer.getTable() |
Table |
LineageInfo.TableAliasInfo.getTable() |
Modifier and Type | Method and Description |
---|---|
void |
LineageInfo.TableAliasInfo.setTable(Table table) |
Constructor and Description |
---|
LineageInfo.DataContainer(Table tab)
Constructor for non partitioned tables.
|
LineageInfo.DataContainer(Table tab,
Partition part)
Constructor for a partitioned tables partition.
|
Modifier and Type | Method and Description |
---|---|
void |
HiveIndexHandler.analyzeIndexDefinition(Table baseTable,
Index index,
Table indexTable)
Requests that the handler validate an index definition and fill in
additional information about its stored representation.
|
void |
AggregateIndexHandler.analyzeIndexDefinition(Table baseTable,
Index index,
Table indexTable) |
Modifier and Type | Method and Description |
---|---|
void |
BitmapIndexHandler.analyzeIndexDefinition(Table baseTable,
Index index,
Table indexTable) |
Modifier and Type | Method and Description |
---|---|
void |
CompactIndexHandler.analyzeIndexDefinition(Table baseTable,
Index index,
Table indexTable) |
Modifier and Type | Method and Description |
---|---|
static Table |
Table.getEmptyTable(String databaseName,
String tableName)
Initialize an empty table.
|
Table |
SessionHiveMetaStoreClient.getTable(String dbname,
String name) |
Table |
Table.getTTable()
This function should only be used in serialization.
|
Modifier and Type | Method and Description |
---|---|
List<Table> |
SessionHiveMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
Modifier and Type | Method and Description |
---|---|
void |
SessionHiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
SessionHiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext envContext) |
protected void |
SessionHiveMetaStoreClient.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
protected void |
Table.initialize(Table table) |
void |
Table.setTTable(Table tTable)
This function should only be called by Java serialization.
|
Constructor and Description |
---|
Table(Table table) |
Modifier and Type | Method and Description |
---|---|
Table |
EximUtil.ReadMetaData.getTable() |
Constructor and Description |
---|
EximUtil.ReadMetaData(Table table,
Iterable<Partition> partitions,
ReplicationSpec replicationSpec) |
Constructor and Description |
---|
AuthorizationPreEventListener.TableWrapper(Table apiTable) |
Modifier and Type | Method and Description |
---|---|
abstract AddPartitionMessage |
MessageFactory.buildAddPartitionMessage(Table table,
Iterator<Partition> partitions)
Factory method for AddPartitionMessage.
|
abstract AlterPartitionMessage |
MessageFactory.buildAlterPartitionMessage(Table table,
Partition before,
Partition after)
Factory method for building AlterPartitionMessage
|
abstract AlterTableMessage |
MessageFactory.buildAlterTableMessage(Table before,
Table after)
Factory method for AlterTableMessage.
|
abstract CreateTableMessage |
MessageFactory.buildCreateTableMessage(Table table)
Factory method for CreateTableMessage.
|
abstract DropPartitionMessage |
MessageFactory.buildDropPartitionMessage(Table table,
Iterator<Partition> partitions)
Factory method for DropPartitionMessage.
|
abstract DropTableMessage |
MessageFactory.buildDropTableMessage(Table table)
Factory method for DropTableMessage.
|
Modifier and Type | Method and Description |
---|---|
AddPartitionMessage |
JSONMessageFactory.buildAddPartitionMessage(Table table,
Iterator<Partition> partitionsIterator) |
AlterPartitionMessage |
JSONMessageFactory.buildAlterPartitionMessage(Table table,
Partition before,
Partition after) |
AlterTableMessage |
JSONMessageFactory.buildAlterTableMessage(Table before,
Table after) |
CreateTableMessage |
JSONMessageFactory.buildCreateTableMessage(Table table) |
DropPartitionMessage |
JSONMessageFactory.buildDropPartitionMessage(Table table,
Iterator<Partition> partitions) |
DropTableMessage |
JSONMessageFactory.buildDropTableMessage(Table table) |
Modifier and Type | Method and Description |
---|---|
protected LazySimpleSerDe |
DelimitedInputWriter.createSerde(Table tbl,
HiveConf conf)
Creates LazySimpleSerde
|
Copyright © 2017 The Apache Software Foundation. All rights reserved.