Package | Description |
---|---|
org.apache.hadoop.hive.accumulo |
Serde and InputFormat support for connecting Hive to Accumulo tables.
|
org.apache.hadoop.hive.hbase |
Implements an HBase storage handler for Hive.
|
org.apache.hadoop.hive.metastore | |
org.apache.hadoop.hive.metastore.api | |
org.apache.hadoop.hive.metastore.events | |
org.apache.hadoop.hive.metastore.parser |
This package implements the parser for parsing the string filter
for the listPartitionsByFilter API.
|
org.apache.hadoop.hive.metastore.partition.spec | |
org.apache.hadoop.hive.metastore.txn | |
org.apache.hadoop.hive.ql.metadata | |
org.apache.hadoop.hive.ql.optimizer.ppr | |
org.apache.hadoop.hive.ql.parse | |
org.apache.hadoop.hive.ql.security.authorization | |
org.apache.hadoop.hive.ql.security.authorization.plugin |
Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements.
|
org.apache.hadoop.hive.ql.txn.compactor | |
org.apache.hive.hcatalog.common | |
org.apache.hive.hcatalog.listener |
Modifier and Type | Method and Description |
---|---|
void |
AccumuloStorageHandler.commitCreateTable(Table table) |
void |
AccumuloStorageHandler.commitDropTable(Table table,
boolean deleteData) |
protected String |
AccumuloStorageHandler.getTableName(Table table) |
void |
AccumuloStorageHandler.preCreateTable(Table table) |
void |
AccumuloStorageHandler.preDropTable(Table table) |
void |
AccumuloStorageHandler.rollbackCreateTable(Table table) |
void |
AccumuloStorageHandler.rollbackDropTable(Table table) |
Modifier and Type | Method and Description |
---|---|
void |
HBaseStorageHandler.commitCreateTable(Table table) |
void |
HBaseStorageHandler.commitDropTable(Table tbl,
boolean deleteData) |
void |
HBaseStorageHandler.preCreateTable(Table tbl) |
void |
HBaseStorageHandler.preDropTable(Table table) |
void |
HBaseStorageHandler.rollbackCreateTable(Table table) |
void |
HBaseStorageHandler.rollbackDropTable(Table table) |
Modifier and Type | Class and Description |
---|---|
static class |
IMetaStoreClient.IncompatibleMetastoreException |
Modifier and Type | Method and Description |
---|---|
Index |
HiveMetaStore.HMSHandler.add_index(Index newIndex,
Table indexTable) |
Partition |
HiveMetaStore.HMSHandler.add_partition_with_environment_context(Partition part,
EnvironmentContext envContext) |
Partition |
IMetaStoreClient.add_partition(Partition partition)
Add a partition to the table.
|
Partition |
HiveMetaStoreClient.add_partition(Partition new_part) |
Partition |
HiveMetaStore.HMSHandler.add_partition(Partition part) |
Partition |
HiveMetaStoreClient.add_partition(Partition new_part,
EnvironmentContext envContext) |
int |
IMetaStoreClient.add_partitions_pspec(PartitionSpecProxy partitionSpec) |
AddPartitionsResult |
HiveMetaStore.HMSHandler.add_partitions_req(AddPartitionsRequest request) |
int |
IMetaStoreClient.add_partitions(List<Partition> partitions)
Add partitions to the table.
|
int |
HiveMetaStoreClient.add_partitions(List<Partition> new_parts) |
int |
HiveMetaStore.HMSHandler.add_partitions(List<Partition> parts) |
List<Partition> |
IMetaStoreClient.add_partitions(List<Partition> partitions,
boolean ifNotExists,
boolean needResults)
Add partitions to the table.
|
List<Partition> |
HiveMetaStoreClient.add_partitions(List<Partition> parts,
boolean ifNotExists,
boolean needResults) |
boolean |
RawStore.addIndex(Index index) |
boolean |
ObjectStore.addIndex(Index index) |
int |
RawStore.addMasterKey(String key) |
int |
ObjectStore.addMasterKey(String key) |
boolean |
RawStore.addPartition(Partition part) |
boolean |
ObjectStore.addPartition(Partition part) |
boolean |
RawStore.addPartitions(String dbName,
String tblName,
List<Partition> parts) |
boolean |
ObjectStore.addPartitions(String dbName,
String tblName,
List<Partition> parts) |
boolean |
RawStore.addPartitions(String dbName,
String tblName,
PartitionSpecProxy partitionSpec,
boolean ifNotExists) |
boolean |
ObjectStore.addPartitions(String dbName,
String tblName,
PartitionSpecProxy partitionSpec,
boolean ifNotExists) |
boolean |
RawStore.addRole(String rowName,
String ownerName) |
boolean |
ObjectStore.addRole(String roleName,
String ownerName) |
void |
HiveMetaStore.HMSHandler.alter_database(String dbName,
Database db) |
void |
HiveMetaStore.HMSHandler.alter_function(String dbName,
String funcName,
Function newFunc) |
void |
IMetaStoreClient.alter_index(String dbName,
String tblName,
String indexName,
Index index) |
void |
HiveMetaStoreClient.alter_index(String dbname,
String base_tbl_name,
String idx_name,
Index new_idx) |
void |
HiveMetaStore.HMSHandler.alter_index(String dbname,
String base_table_name,
String index_name,
Index newIndex) |
void |
HiveMetaStore.HMSHandler.alter_partition_with_environment_context(String dbName,
String tableName,
Partition newPartition,
EnvironmentContext envContext) |
void |
IMetaStoreClient.alter_partition(String dbName,
String tblName,
Partition newPart)
updates a partition to new partition
|
void |
HiveMetaStoreClient.alter_partition(String dbName,
String tblName,
Partition newPart) |
void |
HiveMetaStore.HMSHandler.alter_partition(String db_name,
String tbl_name,
Partition new_part) |
void |
IMetaStoreClient.alter_partitions(String dbName,
String tblName,
List<Partition> newParts)
updates a list of partitions
|
void |
HiveMetaStoreClient.alter_partitions(String dbName,
String tblName,
List<Partition> newParts) |
void |
HiveMetaStore.HMSHandler.alter_partitions(String db_name,
String tbl_name,
List<Partition> new_parts) |
void |
HiveMetaStore.HMSHandler.alter_table_with_cascade(String dbname,
String name,
Table newTable,
boolean cascade) |
void |
HiveMetaStore.HMSHandler.alter_table_with_environment_context(String dbname,
String name,
Table newTable,
EnvironmentContext envContext) |
void |
IMetaStoreClient.alter_table(String defaultDatabaseName,
String tblName,
Table table) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
HiveMetaStore.HMSHandler.alter_table(String dbname,
String name,
Table newTable) |
void |
IMetaStoreClient.alter_table(String defaultDatabaseName,
String tblName,
Table table,
boolean cascade) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
HiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext envContext) |
boolean |
RawStore.alterDatabase(String dbname,
Database db) |
boolean |
ObjectStore.alterDatabase(String dbName,
Database db)
Alter the database object in metastore.
|
void |
IMetaStoreClient.alterDatabase(String name,
Database db) |
void |
HiveMetaStoreClient.alterDatabase(String dbName,
Database db) |
void |
RawStore.alterFunction(String dbName,
String funcName,
Function newFunction)
Alter function based on new function specs.
|
void |
ObjectStore.alterFunction(String dbName,
String funcName,
Function newFunction) |
void |
IMetaStoreClient.alterFunction(String dbName,
String funcName,
Function newFunction) |
void |
HiveMetaStoreClient.alterFunction(String dbName,
String funcName,
Function newFunction) |
void |
RawStore.alterIndex(String dbname,
String baseTblName,
String name,
Index newIndex) |
void |
ObjectStore.alterIndex(String dbname,
String baseTblName,
String name,
Index newIndex) |
Partition |
HiveAlterHandler.alterPartition(RawStore msdb,
Warehouse wh,
String dbname,
String name,
List<String> part_vals,
Partition new_part) |
Partition |
AlterHandler.alterPartition(RawStore msdb,
Warehouse wh,
String dbname,
String name,
List<String> part_vals,
Partition new_part)
handles alter partition
|
void |
RawStore.alterPartition(String db_name,
String tbl_name,
List<String> part_vals,
Partition new_part) |
void |
ObjectStore.alterPartition(String dbname,
String name,
List<String> part_vals,
Partition newPart) |
List<Partition> |
HiveAlterHandler.alterPartitions(RawStore msdb,
Warehouse wh,
String dbname,
String name,
List<Partition> new_parts) |
List<Partition> |
AlterHandler.alterPartitions(RawStore msdb,
Warehouse wh,
String dbname,
String name,
List<Partition> new_part)
handles alter partitions
|
void |
RawStore.alterPartitions(String db_name,
String tbl_name,
List<List<String>> part_vals_list,
List<Partition> new_parts) |
void |
ObjectStore.alterPartitions(String dbname,
String name,
List<List<String>> part_vals,
List<Partition> newParts) |
void |
HiveAlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newt) |
void |
AlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newTable)
handles alter table
|
void |
HiveAlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newt,
boolean cascade) |
void |
AlterHandler.alterTable(RawStore msdb,
Warehouse wh,
String dbname,
String name,
Table newTable,
boolean cascade)
handles alter table, the changes could be cascaded to partitions if applicable
|
void |
RawStore.alterTable(String dbname,
String name,
Table newTable) |
void |
ObjectStore.alterTable(String dbname,
String name,
Table newTable) |
Partition |
HiveMetaStore.HMSHandler.append_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
EnvironmentContext env_context) |
Partition |
HiveMetaStore.HMSHandler.append_partition_by_name(String db_name,
String tbl_name,
String part_name) |
Partition |
HiveMetaStore.HMSHandler.append_partition_with_environment_context(String dbName,
String tableName,
List<String> part_vals,
EnvironmentContext envContext) |
Partition |
HiveMetaStore.HMSHandler.append_partition(String dbName,
String tableName,
List<String> part_vals) |
Partition |
IMetaStoreClient.appendPartition(String tableName,
String dbName,
List<String> partVals) |
Partition |
HiveMetaStoreClient.appendPartition(String db_name,
String table_name,
List<String> part_vals) |
Partition |
HiveMetaStoreClient.appendPartition(String db_name,
String table_name,
List<String> part_vals,
EnvironmentContext envContext) |
Partition |
IMetaStoreClient.appendPartition(String tableName,
String dbName,
String name) |
Partition |
HiveMetaStoreClient.appendPartition(String dbName,
String tableName,
String partName) |
Partition |
HiveMetaStoreClient.appendPartition(String dbName,
String tableName,
String partName,
EnvironmentContext envContext) |
Partition |
HiveMetaStoreClient.appendPartitionByName(String dbName,
String tableName,
String partName) |
Partition |
HiveMetaStoreClient.appendPartitionByName(String dbName,
String tableName,
String partName,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.cancel_delegation_token(String token_str_form) |
void |
IMetaStoreClient.cancelDelegationToken(String tokenStrForm) |
void |
HiveMetaStoreClient.cancelDelegationToken(String tokenStrForm) |
static void |
Deadline.checkTimeout()
Check whether the long running method timeout.
|
static void |
Warehouse.closeFs(org.apache.hadoop.fs.FileSystem fs) |
void |
HiveMetaHook.commitCreateTable(Table table)
Called after successfully adding a new table definition to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.commitDropTable(Table table,
boolean deleteData)
Called after successfully removing a table definition from the metastore
during DROP TABLE.
|
String |
PartitionExpressionProxy.convertExprToFilter(byte[] expr)
Converts serialized Hive expression into filter in the format suitable for Filter.g.
|
static MPartitionColumnStatistics |
StatObjectConverter.convertToMPartitionColumnStatistics(MPartition partition,
ColumnStatisticsDesc statsDesc,
ColumnStatisticsObj statsObj) |
static MTableColumnStatistics |
StatObjectConverter.convertToMTableColumnStatistics(MTable table,
ColumnStatisticsDesc statsDesc,
ColumnStatisticsObj statsObj) |
void |
HiveMetaStore.HMSHandler.create_database(Database db) |
void |
HiveMetaStore.HMSHandler.create_function(Function func) |
boolean |
IMetaStoreClient.create_role(Role role) |
boolean |
HiveMetaStoreClient.create_role(Role role) |
boolean |
HiveMetaStore.HMSHandler.create_role(Role role) |
protected void |
HiveMetaStoreClient.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.create_table(Table tbl) |
boolean |
HiveMetaStore.HMSHandler.create_type(Type type) |
static Table |
MetaStoreUtils.createColumnsetSchema(String name,
List<String> columns,
List<String> partCols,
org.apache.hadoop.conf.Configuration conf) |
void |
RawStore.createDatabase(Database db) |
void |
ObjectStore.createDatabase(Database db) |
void |
IMetaStoreClient.createDatabase(Database db) |
void |
HiveMetaStoreClient.createDatabase(Database db)
Create a new Database
|
void |
RawStore.createFunction(Function func)
Register a user-defined function based on the function specification passed in.
|
void |
ObjectStore.createFunction(Function func) |
void |
IMetaStoreClient.createFunction(Function func) |
void |
HiveMetaStoreClient.createFunction(Function func) |
void |
IMetaStoreClient.createIndex(Index index,
Table indexTable)
create an index
|
void |
HiveMetaStoreClient.createIndex(Index index,
Table indexTable)
create an index
|
void |
RawStore.createTable(Table tbl) |
void |
ObjectStore.createTable(Table tbl) |
void |
IMetaStoreClient.createTable(Table tbl) |
void |
HiveMetaStoreClient.createTable(Table tbl) |
void |
HiveMetaStoreClient.createTable(Table tbl,
EnvironmentContext envContext) |
boolean |
HiveMetaStoreClient.createType(Type type) |
boolean |
HiveMetaStore.HMSHandler.delete_partition_column_statistics(String dbName,
String tableName,
String partName,
String colName) |
boolean |
HiveMetaStore.HMSHandler.delete_table_column_statistics(String dbName,
String tableName,
String colName) |
boolean |
MetaStoreFS.deleteDir(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path f,
boolean recursive,
boolean ifPurge,
org.apache.hadoop.conf.Configuration conf)
delete a directory
|
boolean |
HiveMetaStoreFsImpl.deleteDir(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path f,
boolean recursive,
boolean ifPurge,
org.apache.hadoop.conf.Configuration conf) |
boolean |
Warehouse.deleteDir(org.apache.hadoop.fs.Path f,
boolean recursive) |
boolean |
Warehouse.deleteDir(org.apache.hadoop.fs.Path f,
boolean recursive,
boolean ifPurge) |
boolean |
RawStore.deletePartitionColumnStatistics(String dbName,
String tableName,
String partName,
List<String> partVals,
String colName)
Deletes column statistics if present associated with a given db, table, partition and col.
|
boolean |
ObjectStore.deletePartitionColumnStatistics(String dbName,
String tableName,
String partName,
List<String> partVals,
String colName) |
boolean |
IMetaStoreClient.deletePartitionColumnStatistics(String dbName,
String tableName,
String partName,
String colName)
Delete partition level column statistics given dbName, tableName, partName and colName
|
boolean |
HiveMetaStoreClient.deletePartitionColumnStatistics(String dbName,
String tableName,
String partName,
String colName)
Delete partition level column statistics given dbName, tableName, partName and colName
|
boolean |
RawStore.deleteTableColumnStatistics(String dbName,
String tableName,
String colName)
Deletes column statistics if present associated with a given db, table and col.
|
boolean |
ObjectStore.deleteTableColumnStatistics(String dbName,
String tableName,
String colName) |
boolean |
IMetaStoreClient.deleteTableColumnStatistics(String dbName,
String tableName,
String colName)
Delete table level column statistics given dbName, tableName and colName
|
boolean |
HiveMetaStoreClient.deleteTableColumnStatistics(String dbName,
String tableName,
String colName)
Delete table level column statistics given dbName, tableName and colName
|
static void |
MetaStoreUtils.deleteWHDirectory(org.apache.hadoop.fs.Path path,
org.apache.hadoop.conf.Configuration conf,
boolean use_trash) |
boolean |
RawStore.doesPartitionExist(String dbName,
String tableName,
List<String> part_vals) |
boolean |
ObjectStore.doesPartitionExist(String dbName,
String tableName,
List<String> partVals) |
void |
HiveMetaStore.HMSHandler.drop_database(String dbName,
boolean deleteData,
boolean cascade) |
void |
HiveMetaStore.HMSHandler.drop_function(String dbName,
String funcName) |
boolean |
HiveMetaStore.HMSHandler.drop_index_by_name(String dbName,
String tblName,
String indexName,
boolean deleteData) |
boolean |
HiveMetaStore.HMSHandler.drop_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
boolean deleteData,
EnvironmentContext envContext) |
boolean |
HiveMetaStore.HMSHandler.drop_partition_by_name(String db_name,
String tbl_name,
String part_name,
boolean deleteData) |
boolean |
HiveMetaStore.HMSHandler.drop_partition_with_environment_context(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData,
EnvironmentContext envContext) |
boolean |
HiveMetaStore.HMSHandler.drop_partition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData) |
DropPartitionsResult |
HiveMetaStore.HMSHandler.drop_partitions_req(DropPartitionsRequest request) |
boolean |
IMetaStoreClient.drop_role(String role_name) |
boolean |
HiveMetaStoreClient.drop_role(String roleName) |
boolean |
HiveMetaStore.HMSHandler.drop_role(String roleName) |
protected void |
HiveMetaStoreClient.drop_table_with_environment_context(String dbname,
String name,
boolean deleteData,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.drop_table_with_environment_context(String dbname,
String name,
boolean deleteData,
EnvironmentContext envContext) |
void |
HiveMetaStore.HMSHandler.drop_table(String dbname,
String name,
boolean deleteData) |
boolean |
HiveMetaStore.HMSHandler.drop_type(String name) |
boolean |
RawStore.dropDatabase(String dbname) |
boolean |
ObjectStore.dropDatabase(String dbname) |
void |
IMetaStoreClient.dropDatabase(String name) |
void |
HiveMetaStoreClient.dropDatabase(String name) |
void |
IMetaStoreClient.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb) |
void |
HiveMetaStoreClient.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb) |
void |
IMetaStoreClient.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb,
boolean cascade) |
void |
HiveMetaStoreClient.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb,
boolean cascade) |
void |
RawStore.dropFunction(String dbName,
String funcName)
Drop a function definition.
|
void |
ObjectStore.dropFunction(String dbName,
String funcName) |
void |
IMetaStoreClient.dropFunction(String dbName,
String funcName) |
void |
HiveMetaStoreClient.dropFunction(String dbName,
String funcName) |
boolean |
RawStore.dropIndex(String dbName,
String origTableName,
String indexName) |
boolean |
ObjectStore.dropIndex(String dbName,
String origTableName,
String indexName) |
boolean |
IMetaStoreClient.dropIndex(String db_name,
String tbl_name,
String name,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropIndex(String dbName,
String tblName,
String name,
boolean deleteData) |
boolean |
RawStore.dropPartition(String dbName,
String tableName,
List<String> part_vals) |
boolean |
ObjectStore.dropPartition(String dbName,
String tableName,
List<String> part_vals) |
boolean |
HiveMetaStoreClient.dropPartition(String db_name,
String tbl_name,
List<String> part_vals) |
boolean |
IMetaStoreClient.dropPartition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropPartition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropPartition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData,
EnvironmentContext envContext) |
boolean |
HiveMetaStoreClient.dropPartition(String db_name,
String tbl_name,
List<String> part_vals,
EnvironmentContext env_context) |
boolean |
IMetaStoreClient.dropPartition(String db_name,
String tbl_name,
String name,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropPartition(String dbName,
String tableName,
String partName,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropPartition(String dbName,
String tableName,
String partName,
boolean deleteData,
EnvironmentContext envContext) |
boolean |
HiveMetaStoreClient.dropPartitionByName(String dbName,
String tableName,
String partName,
boolean deleteData) |
boolean |
HiveMetaStoreClient.dropPartitionByName(String dbName,
String tableName,
String partName,
boolean deleteData,
EnvironmentContext envContext) |
List<Partition> |
IMetaStoreClient.dropPartitions(String dbName,
String tblName,
List<ObjectPair<Integer,byte[]>> partExprs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists) |
List<Partition> |
HiveMetaStoreClient.dropPartitions(String dbName,
String tblName,
List<ObjectPair<Integer,byte[]>> partExprs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists) |
List<Partition> |
IMetaStoreClient.dropPartitions(String dbName,
String tblName,
List<ObjectPair<Integer,byte[]>> partExprs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists,
boolean needResults) |
List<Partition> |
HiveMetaStoreClient.dropPartitions(String dbName,
String tblName,
List<ObjectPair<Integer,byte[]>> partExprs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists,
boolean needResult) |
void |
RawStore.dropPartitions(String dbName,
String tblName,
List<String> partNames) |
void |
ObjectStore.dropPartitions(String dbName,
String tblName,
List<String> partNames) |
void |
IMetaStoreClient.dropTable(String tableName,
boolean deleteData)
Deprecated.
As of release 0.6.0 replaced by
IMetaStoreClient.dropTable(String, String, boolean, boolean) .
This method will be removed in release 0.7.0. |
void |
HiveMetaStoreClient.dropTable(String tableName,
boolean deleteData)
Deprecated.
|
boolean |
RawStore.dropTable(String dbName,
String tableName) |
boolean |
ObjectStore.dropTable(String dbName,
String tableName) |
void |
IMetaStoreClient.dropTable(String dbname,
String tableName) |
void |
HiveMetaStoreClient.dropTable(String dbname,
String name) |
void |
IMetaStoreClient.dropTable(String dbname,
String tableName,
boolean deleteData,
boolean ignoreUnknownTab)
Drop the table.
|
void |
HiveMetaStoreClient.dropTable(String dbname,
String name,
boolean deleteData,
boolean ignoreUnknownTab)
Drop the table.
|
void |
IMetaStoreClient.dropTable(String dbname,
String tableName,
boolean deleteData,
boolean ignoreUnknownTab,
boolean ifPurge) |
void |
HiveMetaStoreClient.dropTable(String dbname,
String name,
boolean deleteData,
boolean ignoreUnknownTab,
boolean ifPurge)
Drop the table and choose whether to save the data in the trash.
|
void |
HiveMetaStoreClient.dropTable(String dbname,
String name,
boolean deleteData,
boolean ignoreUnknownTab,
EnvironmentContext envContext)
Drop the table and choose whether to: delete the underlying table data;
throw if the table doesn't exist; save the data in the trash.
|
boolean |
HiveMetaStoreClient.dropType(String type) |
Partition |
IMetaStoreClient.exchange_partition(Map<String,String> partitionSpecs,
String sourceDb,
String sourceTable,
String destdb,
String destTableName) |
Partition |
HiveMetaStoreClient.exchange_partition(Map<String,String> partitionSpecs,
String sourceDb,
String sourceTable,
String destDb,
String destinationTableName)
Exchange the partition between two tables
|
Partition |
HiveMetaStore.HMSHandler.exchange_partition(Map<String,String> partitionSpecs,
String sourceDbName,
String sourceTableName,
String destDbName,
String destTableName) |
static void |
StatObjectConverter.fillColumnStatisticsData(String colType,
ColumnStatisticsData data,
Object llow,
Object lhigh,
Object dlow,
Object dhigh,
Object declow,
Object dechigh,
Object nulls,
Object dist,
Object avglen,
Object maxlen,
Object trues,
Object falses) |
static void |
StatObjectConverter.fillColumnStatisticsData(String colType,
ColumnStatisticsData data,
Object llow,
Object lhigh,
Object dlow,
Object dhigh,
Object declow,
Object dechigh,
Object nulls,
Object dist,
Object avglen,
Object maxlen,
Object trues,
Object falses,
Object avgLong,
Object avgDouble,
Object avgDecimal,
Object sumDist,
boolean useDensityFunctionForNDVEstimation) |
Database |
MetaStoreFilterHook.filterDatabase(Database dataBase)
filter to given database object if applicable
|
List<String> |
MetaStoreFilterHook.filterDatabases(List<String> dbList)
Filter given list of databases
|
List<String> |
DefaultMetaStoreFilterHookImpl.filterDatabases(List<String> dbList) |
Index |
MetaStoreFilterHook.filterIndex(Index index) |
List<Index> |
MetaStoreFilterHook.filterIndexes(List<Index> indexeList)
Filter given list of index objects
|
List<Index> |
DefaultMetaStoreFilterHookImpl.filterIndexes(List<Index> indexeList) |
List<String> |
MetaStoreFilterHook.filterIndexNames(String dbName,
String tblName,
List<String> indexList)
Filter given list of index names
|
List<String> |
DefaultMetaStoreFilterHookImpl.filterIndexNames(String dbName,
String tblName,
List<String> indexList) |
Partition |
MetaStoreFilterHook.filterPartition(Partition partition)
filter to given partition object if applicable
|
List<String> |
MetaStoreFilterHook.filterPartitionNames(String dbName,
String tblName,
List<String> partitionNames)
Filter given list of partition names
|
List<String> |
DefaultMetaStoreFilterHookImpl.filterPartitionNames(String dbName,
String tblName,
List<String> partitionNames) |
List<Partition> |
MetaStoreFilterHook.filterPartitions(List<Partition> partitionList)
Filter given list of partitions
|
List<Partition> |
DefaultMetaStoreFilterHookImpl.filterPartitions(List<Partition> partitionList) |
boolean |
PartitionExpressionProxy.filterPartitionsByExpr(List<String> partColumnNames,
List<PrimitiveTypeInfo> partColumnTypeInfos,
byte[] expr,
String defaultPartitionName,
List<String> partitionNames)
Filters the partition names via serialized Hive expression.
|
List<PartitionSpec> |
MetaStoreFilterHook.filterPartitionSpecs(List<PartitionSpec> partitionSpecList)
Filter given list of partition specs
|
List<PartitionSpec> |
DefaultMetaStoreFilterHookImpl.filterPartitionSpecs(List<PartitionSpec> partitionSpecList) |
Table |
MetaStoreFilterHook.filterTable(Table table)
filter to given table object if applicable
|
List<String> |
MetaStoreFilterHook.filterTableNames(String dbName,
List<String> tableList)
Filter given list of tables
|
List<String> |
DefaultMetaStoreFilterHookImpl.filterTableNames(String dbName,
List<String> tableList) |
List<Table> |
MetaStoreFilterHook.filterTables(List<Table> tableList)
Filter given list of tables
|
List<Table> |
DefaultMetaStoreFilterHookImpl.filterTables(List<Table> tableList) |
AggrStats |
HiveMetaStore.HMSHandler.get_aggr_stats_for(PartitionsStatsRequest request) |
AggrStats |
RawStore.get_aggr_stats_for(String dbName,
String tblName,
List<String> partNames,
List<String> colNames) |
AggrStats |
ObjectStore.get_aggr_stats_for(String dbName,
String tblName,
List<String> partNames,
List<String> colNames) |
List<String> |
HiveMetaStore.HMSHandler.get_all_databases() |
List<String> |
HiveMetaStore.HMSHandler.get_all_tables(String dbname) |
Database |
HiveMetaStore.HMSHandler.get_database_core(String name)
Equivalent to get_database, but does not write to audit logs, or fire pre-event listners.
|
Database |
HiveMetaStore.HMSHandler.get_database(String name) |
List<String> |
HiveMetaStore.HMSHandler.get_databases(String pattern) |
String |
HiveMetaStore.HMSHandler.get_delegation_token(String token_owner,
String renewer_kerberos_principal_name) |
List<FieldSchema> |
HiveMetaStore.HMSHandler.get_fields_with_environment_context(String db,
String tableName,
EnvironmentContext envContext) |
List<FieldSchema> |
HiveMetaStore.HMSHandler.get_fields(String db,
String tableName) |
Function |
HiveMetaStore.HMSHandler.get_function(String dbName,
String funcName) |
List<String> |
HiveMetaStore.HMSHandler.get_functions(String dbName,
String pattern) |
Index |
HiveMetaStore.HMSHandler.get_index_by_name(String dbName,
String tblName,
String indexName) |
List<String> |
HiveMetaStore.HMSHandler.get_index_names(String dbName,
String tblName,
short maxIndexes) |
List<Index> |
HiveMetaStore.HMSHandler.get_indexes(String dbName,
String tblName,
short maxIndexes) |
List<PartitionSpec> |
HiveMetaStore.HMSHandler.get_part_specs_by_filter(String dbName,
String tblName,
String filter,
int maxParts) |
Partition |
HiveMetaStore.HMSHandler.get_partition_by_name(String db_name,
String tbl_name,
String part_name) |
ColumnStatistics |
HiveMetaStore.HMSHandler.get_partition_column_statistics(String dbName,
String tableName,
String partName,
String colName) |
List<String> |
HiveMetaStore.HMSHandler.get_partition_names_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<String> |
HiveMetaStore.HMSHandler.get_partition_names(String db_name,
String tbl_name,
short max_parts) |
Partition |
HiveMetaStore.HMSHandler.get_partition_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
String user_name,
List<String> group_names) |
Partition |
HiveMetaStore.HMSHandler.get_partition(String db_name,
String tbl_name,
List<String> part_vals) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions_by_filter(String dbName,
String tblName,
String filter,
short maxParts) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions_by_names(String dbName,
String tblName,
List<String> partNames) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions_ps_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String userName,
List<String> groupNames) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<PartitionSpec> |
HiveMetaStore.HMSHandler.get_partitions_pspec(String db_name,
String tbl_name,
int max_parts) |
PartitionsStatsResult |
HiveMetaStore.HMSHandler.get_partitions_statistics_req(PartitionsStatsRequest request) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions_with_auth(String dbName,
String tblName,
short maxParts,
String userName,
List<String> groupNames) |
List<Partition> |
HiveMetaStore.HMSHandler.get_partitions(String db_name,
String tbl_name,
short max_parts) |
GetPrincipalsInRoleResponse |
IMetaStoreClient.get_principals_in_role(GetPrincipalsInRoleRequest getPrincRoleReq)
get all role-grants for users/roles that have been granted the given role
Note that in the returned list of RolePrincipalGrants, the roleName is
redundant as it would match the role_name argument of this function
|
GetPrincipalsInRoleResponse |
HiveMetaStoreClient.get_principals_in_role(GetPrincipalsInRoleRequest req) |
GetPrincipalsInRoleResponse |
HiveMetaStore.HMSHandler.get_principals_in_role(GetPrincipalsInRoleRequest request) |
PrincipalPrivilegeSet |
IMetaStoreClient.get_privilege_set(HiveObjectRef hiveObject,
String user_name,
List<String> group_names)
Return the privileges that the user, group have directly and indirectly through roles
on the given hiveObject
|
PrincipalPrivilegeSet |
HiveMetaStoreClient.get_privilege_set(HiveObjectRef hiveObject,
String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
HiveMetaStore.HMSHandler.get_privilege_set(HiveObjectRef hiveObject,
String userName,
List<String> groupNames) |
GetRoleGrantsForPrincipalResponse |
IMetaStoreClient.get_role_grants_for_principal(GetRoleGrantsForPrincipalRequest getRolePrincReq)
get all role-grants for roles that have been granted to given principal
Note that in the returned list of RolePrincipalGrants, the principal information
redundant as it would match the principal information in request
|
GetRoleGrantsForPrincipalResponse |
HiveMetaStoreClient.get_role_grants_for_principal(GetRoleGrantsForPrincipalRequest getRolePrincReq) |
GetRoleGrantsForPrincipalResponse |
HiveMetaStore.HMSHandler.get_role_grants_for_principal(GetRoleGrantsForPrincipalRequest request) |
List<String> |
HiveMetaStore.HMSHandler.get_role_names() |
List<FieldSchema> |
HiveMetaStore.HMSHandler.get_schema_with_environment_context(String db,
String tableName,
EnvironmentContext envContext)
Return the schema of the table.
|
List<FieldSchema> |
HiveMetaStore.HMSHandler.get_schema(String db,
String tableName)
Return the schema of the table.
|
ColumnStatistics |
HiveMetaStore.HMSHandler.get_table_column_statistics(String dbName,
String tableName,
String colName) |
Table |
HiveMetaStore.HMSHandler.get_table_core(String dbname,
String name)
Equivalent of get_table, but does not log audits and fire pre-event listener.
|
List<String> |
HiveMetaStore.HMSHandler.get_table_names_by_filter(String dbName,
String filter,
short maxTables) |
List<Table> |
HiveMetaStore.HMSHandler.get_table_objects_by_name(String dbname,
List<String> names)
Gets multiple tables from the hive metastore.
|
TableStatsResult |
HiveMetaStore.HMSHandler.get_table_statistics_req(TableStatsRequest request) |
Table |
HiveMetaStore.HMSHandler.get_table(String dbname,
String name) |
List<String> |
HiveMetaStore.HMSHandler.get_tables(String dbname,
String pattern) |
Map<String,Type> |
HiveMetaStore.HMSHandler.get_type_all(String name) |
Type |
HiveMetaStore.HMSHandler.get_type(String name) |
AggrStats |
IMetaStoreClient.getAggrColStatsFor(String dbName,
String tblName,
List<String> colNames,
List<String> partName) |
AggrStats |
HiveMetaStoreClient.getAggrColStatsFor(String dbName,
String tblName,
List<String> colNames,
List<String> partNames) |
List<String> |
RawStore.getAllDatabases() |
List<String> |
ObjectStore.getAllDatabases() |
List<String> |
IMetaStoreClient.getAllDatabases()
Get the names of all databases in the MetaStore.
|
List<String> |
HiveMetaStoreClient.getAllDatabases()
Get the names of all databases in the MetaStore.
|
List<String> |
RawStore.getAllTables(String dbName) |
List<String> |
ObjectStore.getAllTables(String dbName) |
List<String> |
IMetaStoreClient.getAllTables(String dbName)
Get the names of all tables in the specified database.
|
List<String> |
HiveMetaStoreClient.getAllTables(String dbname)
Get the names of all tables in the specified database.
|
static int |
MetaStoreUtils.getArchivingLevel(Partition part) |
static Class<?> |
MetaStoreUtils.getClass(String rawStoreClassName) |
PrincipalPrivilegeSet |
RawStore.getColumnPrivilegeSet(String dbName,
String tableName,
String partitionName,
String columnName,
String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
ObjectStore.getColumnPrivilegeSet(String dbName,
String tableName,
String partitionName,
String columnName,
String userName,
List<String> groupNames) |
Database |
IMetaStoreClient.getDatabase(String databaseName)
Get a Database Object
|
Database |
HiveMetaStoreClient.getDatabase(String name) |
Database |
ObjectStore.getDatabaseInternal(String name) |
org.apache.hadoop.fs.Path |
Warehouse.getDatabasePath(Database db) |
List<String> |
RawStore.getDatabases(String pattern) |
List<String> |
ObjectStore.getDatabases(String pattern) |
List<String> |
IMetaStoreClient.getDatabases(String databasePattern)
Get the names of all databases in the MetaStore that match the given pattern.
|
List<String> |
HiveMetaStoreClient.getDatabases(String databasePattern)
Get the names of all databases in the MetaStore that match the given pattern.
|
List<PrivilegeGrantInfo> |
ObjectStore.getDBPrivilege(String dbName,
String principalName,
PrincipalType principalType) |
PrincipalPrivilegeSet |
RawStore.getDBPrivilegeSet(String dbName,
String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
ObjectStore.getDBPrivilegeSet(String dbName,
String userName,
List<String> groupNames) |
org.apache.hadoop.fs.Path |
Warehouse.getDefaultDatabasePath(String dbName) |
String |
HiveMetaStoreClient.getDelegationToken(String renewerKerberosPrincipalName) |
String |
IMetaStoreClient.getDelegationToken(String owner,
String renewerKerberosPrincipalName)
This is expected to be a no-op when in local mode,
which means that the implementation will return null.
|
String |
HiveMetaStoreClient.getDelegationToken(String owner,
String renewerKerberosPrincipalName) |
static Deserializer |
MetaStoreUtils.getDeserializer(org.apache.hadoop.conf.Configuration conf,
Partition part,
Table table)
getDeserializer
Get the Deserializer for a partition.
|
static Deserializer |
MetaStoreUtils.getDeserializer(org.apache.hadoop.conf.Configuration conf,
Table table,
boolean skipConfError)
getDeserializer
Get the Deserializer for a table.
|
org.apache.hadoop.fs.Path |
Warehouse.getDnsPath(org.apache.hadoop.fs.Path path) |
static org.apache.hadoop.fs.Path |
Warehouse.getDnsPath(org.apache.hadoop.fs.Path path,
org.apache.hadoop.conf.Configuration conf)
Hadoop File System reverse lookups paths with raw ip addresses The File
System URI always contains the canonical DNS name of the Namenode.
|
List<FieldSchema> |
IMetaStoreClient.getFields(String db,
String tableName) |
List<FieldSchema> |
HiveMetaStoreClient.getFields(String db,
String tableName) |
static List<FieldSchema> |
MetaStoreUtils.getFieldsFromDeserializer(String tableName,
Deserializer deserializer) |
org.apache.hadoop.fs.FileStatus[] |
Warehouse.getFileStatusesForLocation(String location) |
org.apache.hadoop.fs.FileStatus[] |
Warehouse.getFileStatusesForSD(StorageDescriptor desc) |
org.apache.hadoop.fs.FileStatus[] |
Warehouse.getFileStatusesForUnpartitionedTable(Database db,
Table table) |
org.apache.hadoop.fs.FileSystem |
Warehouse.getFs(org.apache.hadoop.fs.Path f) |
static org.apache.hadoop.fs.FileSystem |
Warehouse.getFs(org.apache.hadoop.fs.Path f,
org.apache.hadoop.conf.Configuration conf)
Helper functions to convert IOException to MetaException
|
Function |
RawStore.getFunction(String dbName,
String funcName)
Retrieve function by name.
|
Function |
ObjectStore.getFunction(String dbName,
String funcName) |
Function |
IMetaStoreClient.getFunction(String dbName,
String funcName) |
Function |
HiveMetaStoreClient.getFunction(String dbName,
String funcName) |
List<String> |
RawStore.getFunctions(String dbName,
String pattern)
Retrieve list of function names based on name pattern.
|
List<String> |
ObjectStore.getFunctions(String dbName,
String pattern) |
List<String> |
IMetaStoreClient.getFunctions(String dbName,
String pattern) |
List<String> |
HiveMetaStoreClient.getFunctions(String dbName,
String pattern) |
HiveMetaHook |
HiveMetaHookLoader.getHook(Table tbl)
Loads a hook for the specified table.
|
Index |
RawStore.getIndex(String dbName,
String origTableName,
String indexName) |
Index |
ObjectStore.getIndex(String dbName,
String origTableName,
String indexName) |
Index |
IMetaStoreClient.getIndex(String dbName,
String tblName,
String indexName) |
Index |
HiveMetaStoreClient.getIndex(String dbName,
String tblName,
String indexName) |
List<Index> |
RawStore.getIndexes(String dbName,
String origTableName,
int max) |
List<Index> |
ObjectStore.getIndexes(String dbName,
String origTableName,
int max) |
String |
IMetaStoreClient.getMetaConf(String key)
get current meta variable
|
String |
HiveMetaStore.HMSHandler.getMetaConf(String key) |
String |
RawStore.getMetaStoreSchemaVersion() |
String |
ObjectStore.getMetaStoreSchemaVersion() |
RawStore |
HiveMetaStore.HMSHandler.getMS()
Get a cached RawStore.
|
Partition |
RawStore.getPartition(String dbName,
String tableName,
List<String> part_vals) |
Partition |
ObjectStore.getPartition(String dbName,
String tableName,
List<String> part_vals) |
Partition |
IMetaStoreClient.getPartition(String tblName,
String dbName,
List<String> partVals) |
Partition |
HiveMetaStoreClient.getPartition(String db_name,
String tbl_name,
List<String> part_vals) |
Partition |
IMetaStoreClient.getPartition(String dbName,
String tblName,
String name) |
Partition |
HiveMetaStoreClient.getPartition(String db,
String tableName,
String partName) |
List<ColumnStatistics> |
RawStore.getPartitionColumnStatistics(String dbName,
String tblName,
List<String> partNames,
List<String> colNames)
Returns the relevant column statistics for given columns in given partitions in a given
table in a given database if such statistics exist.
|
List<ColumnStatistics> |
ObjectStore.getPartitionColumnStatistics(String dbName,
String tableName,
List<String> partNames,
List<String> colNames) |
Map<String,List<ColumnStatisticsObj>> |
IMetaStoreClient.getPartitionColumnStatistics(String dbName,
String tableName,
List<String> partNames,
List<String> colNames)
Get partitions column statistics given dbName, tableName, multiple partitions and colName-s
|
Map<String,List<ColumnStatisticsObj>> |
HiveMetaStoreClient.getPartitionColumnStatistics(String dbName,
String tableName,
List<String> partNames,
List<String> colNames)
Get partitions column statistics given dbName, tableName, multiple partitions and colName-s
|
protected List<ColumnStatistics> |
ObjectStore.getPartitionColumnStatisticsInternal(String dbName,
String tableName,
List<String> partNames,
List<String> colNames,
boolean allowSql,
boolean allowJdo) |
org.apache.hadoop.fs.Path |
Warehouse.getPartitionPath(Database db,
String tableName,
LinkedHashMap<String,String> pm) |
org.apache.hadoop.fs.Path |
Warehouse.getPartitionPath(org.apache.hadoop.fs.Path tblPath,
LinkedHashMap<String,String> pm) |
PrincipalPrivilegeSet |
RawStore.getPartitionPrivilegeSet(String dbName,
String tableName,
String partition,
String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
ObjectStore.getPartitionPrivilegeSet(String dbName,
String tableName,
String partition,
String userName,
List<String> groupNames) |
List<Partition> |
RawStore.getPartitions(String dbName,
String tableName,
int max) |
List<Partition> |
ObjectStore.getPartitions(String dbName,
String tableName,
int maxParts) |
List<Partition> |
RawStore.getPartitionsByFilter(String dbName,
String tblName,
String filter,
short maxParts) |
List<Partition> |
ObjectStore.getPartitionsByFilter(String dbName,
String tblName,
String filter,
short maxParts) |
protected List<Partition> |
ObjectStore.getPartitionsByFilterInternal(String dbName,
String tblName,
String filter,
short maxParts,
boolean allowSql,
boolean allowJdo) |
List<Partition> |
RawStore.getPartitionsByNames(String dbName,
String tblName,
List<String> partNames) |
List<Partition> |
ObjectStore.getPartitionsByNames(String dbName,
String tblName,
List<String> partNames) |
List<Partition> |
IMetaStoreClient.getPartitionsByNames(String db_name,
String tbl_name,
List<String> part_names)
Get partitions by a list of partition names.
|
List<Partition> |
HiveMetaStoreClient.getPartitionsByNames(String db_name,
String tbl_name,
List<String> part_names) |
protected List<Partition> |
ObjectStore.getPartitionsByNamesInternal(String dbName,
String tblName,
List<String> partNames,
boolean allowSql,
boolean allowJdo) |
protected List<Partition> |
ObjectStore.getPartitionsInternal(String dbName,
String tblName,
int maxParts,
boolean allowSql,
boolean allowJdo) |
List<Partition> |
RawStore.getPartitionsWithAuth(String dbName,
String tblName,
short maxParts,
String userName,
List<String> groupNames) |
List<Partition> |
ObjectStore.getPartitionsWithAuth(String dbName,
String tblName,
short max,
String userName,
List<String> groupNames) |
Partition |
RawStore.getPartitionWithAuth(String dbName,
String tblName,
List<String> partVals,
String user_name,
List<String> group_names) |
Partition |
ObjectStore.getPartitionWithAuth(String dbName,
String tblName,
List<String> partVals,
String user_name,
List<String> group_names) |
Partition |
IMetaStoreClient.getPartitionWithAuthInfo(String dbName,
String tableName,
List<String> pvals,
String userName,
List<String> groupNames) |
Partition |
HiveMetaStoreClient.getPartitionWithAuthInfo(String db_name,
String tbl_name,
List<String> part_vals,
String user_name,
List<String> group_names) |
static List<String> |
Warehouse.getPartValuesFromPartName(String partName) |
static IMetaStoreClient |
RetryingMetaStoreClient.getProxy(HiveConf hiveConf) |
static IMetaStoreClient |
RetryingMetaStoreClient.getProxy(HiveConf hiveConf,
Class<?>[] constructorArgTypes,
Object[] constructorArgs,
Map<String,Long> metaCallTimeMap,
String mscClassName)
This constructor is meant for Hive internal use only.
|
static IMetaStoreClient |
RetryingMetaStoreClient.getProxy(HiveConf hiveConf,
Class<?>[] constructorArgTypes,
Object[] constructorArgs,
String mscClassName)
This constructor is meant for Hive internal use only.
|
static RawStore |
RawStoreProxy.getProxy(HiveConf hiveConf,
org.apache.hadoop.conf.Configuration conf,
String rawStoreClassName,
int id) |
static IMetaStoreClient |
RetryingMetaStoreClient.getProxy(HiveConf hiveConf,
HiveMetaHookLoader hookLoader,
Map<String,Long> metaCallTimeMap,
String mscClassName) |
static IMetaStoreClient |
RetryingMetaStoreClient.getProxy(HiveConf hiveConf,
HiveMetaHookLoader hookLoader,
String mscClassName) |
static IHMSHandler |
RetryingHMSHandler.getProxy(HiveConf hiveConf,
IHMSHandler baseHandler,
boolean local) |
List<FieldSchema> |
IMetaStoreClient.getSchema(String db,
String tableName) |
List<FieldSchema> |
HiveMetaStoreClient.getSchema(String db,
String tableName) |
Table |
IMetaStoreClient.getTable(String tableName)
Deprecated.
As of release 0.6.0 replaced by
IMetaStoreClient.getTable(String, String) .
This method will be removed in release 0.7.0. |
Table |
HiveMetaStoreClient.getTable(String tableName)
Deprecated.
|
Table |
RawStore.getTable(String dbName,
String tableName) |
Table |
ObjectStore.getTable(String dbName,
String tableName) |
Table |
IMetaStoreClient.getTable(String dbName,
String tableName)
Get a table object.
|
Table |
HiveMetaStoreClient.getTable(String dbname,
String name) |
ColumnStatistics |
RawStore.getTableColumnStatistics(String dbName,
String tableName,
List<String> colName)
Returns the relevant column statistics for a given column in a given table in a given database
if such statistics exist.
|
ColumnStatistics |
ObjectStore.getTableColumnStatistics(String dbName,
String tableName,
List<String> colNames) |
List<ColumnStatisticsObj> |
IMetaStoreClient.getTableColumnStatistics(String dbName,
String tableName,
List<String> colNames)
Get table column statistics given dbName, tableName and multiple colName-s
|
List<ColumnStatisticsObj> |
HiveMetaStoreClient.getTableColumnStatistics(String dbName,
String tableName,
List<String> colNames)
Get table column statistics given dbName, tableName and multiple colName-s
|
protected ColumnStatistics |
ObjectStore.getTableColumnStatisticsInternal(String dbName,
String tableName,
List<String> colNames,
boolean allowSql,
boolean allowJdo) |
List<Table> |
RawStore.getTableObjectsByName(String dbname,
List<String> tableNames) |
List<Table> |
ObjectStore.getTableObjectsByName(String db,
List<String> tbl_names) |
List<Table> |
IMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
List<Table> |
HiveMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
org.apache.hadoop.fs.Path |
Warehouse.getTablePath(Database db,
String tableName) |
org.apache.hadoop.fs.Path |
Warehouse.getTablePath(String whRootString,
String tableName) |
PrincipalPrivilegeSet |
RawStore.getTablePrivilegeSet(String dbName,
String tableName,
String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
ObjectStore.getTablePrivilegeSet(String dbName,
String tableName,
String userName,
List<String> groupNames) |
List<String> |
RawStore.getTables(String dbName,
String pattern) |
List<String> |
ObjectStore.getTables(String dbName,
String pattern) |
List<String> |
IMetaStoreClient.getTables(String dbName,
String tablePattern)
Get the names of all tables in the specified database that satisfy the supplied
table name pattern.
|
List<String> |
HiveMetaStoreClient.getTables(String dbname,
String tablePattern)
Get the names of all tables in the specified database that satisfy the supplied
table name pattern.
|
Type |
HiveMetaStoreClient.getType(String name) |
Map<String,Type> |
HiveMetaStoreClient.getTypeAll(String name) |
PrincipalPrivilegeSet |
RawStore.getUserPrivilegeSet(String userName,
List<String> groupNames) |
PrincipalPrivilegeSet |
ObjectStore.getUserPrivilegeSet(String userName,
List<String> groupNames) |
org.apache.hadoop.fs.Path |
Warehouse.getWhRoot()
Resolve the configured warehouse root dir with respect to the configuration
This involves opening the FileSystem corresponding to the warehouse root
dir (but that should be ok given that this is only called during DDL
statements for non-external tables).
|
boolean |
IMetaStoreClient.grant_privileges(PrivilegeBag privileges) |
boolean |
HiveMetaStoreClient.grant_privileges(PrivilegeBag privileges) |
boolean |
HiveMetaStore.HMSHandler.grant_privileges(PrivilegeBag privileges) |
GrantRevokePrivilegeResponse |
HiveMetaStore.HMSHandler.grant_revoke_privileges(GrantRevokePrivilegeRequest request) |
GrantRevokeRoleResponse |
HiveMetaStore.HMSHandler.grant_revoke_role(GrantRevokeRoleRequest request) |
boolean |
IMetaStoreClient.grant_role(String role_name,
String user_name,
PrincipalType principalType,
String grantor,
PrincipalType grantorType,
boolean grantOption) |
boolean |
HiveMetaStoreClient.grant_role(String roleName,
String userName,
PrincipalType principalType,
String grantor,
PrincipalType grantorType,
boolean grantOption) |
boolean |
HiveMetaStore.HMSHandler.grant_role(String roleName,
String principalName,
PrincipalType principalType,
String grantor,
PrincipalType grantorType,
boolean grantOption) |
boolean |
RawStore.grantPrivileges(PrivilegeBag privileges) |
boolean |
ObjectStore.grantPrivileges(PrivilegeBag privileges) |
boolean |
RawStore.grantRole(Role role,
String userName,
PrincipalType principalType,
String grantor,
PrincipalType grantorType,
boolean grantOption) |
boolean |
ObjectStore.grantRole(Role role,
String userName,
PrincipalType principalType,
String grantor,
PrincipalType grantorType,
boolean grantOption) |
void |
IHMSHandler.init() |
void |
HiveMetaStore.HMSHandler.init() |
void |
MetaStoreThread.init(AtomicBoolean stop,
AtomicBoolean looped)
Initialize the thread.
|
boolean |
Warehouse.isDir(org.apache.hadoop.fs.Path f) |
boolean |
Warehouse.isEmpty(org.apache.hadoop.fs.Path path) |
boolean |
RawStore.isPartitionMarkedForEvent(String dbName,
String tblName,
Map<String,String> partName,
PartitionEventType evtType) |
boolean |
ObjectStore.isPartitionMarkedForEvent(String dbName,
String tblName,
Map<String,String> partName,
PartitionEventType evtType) |
boolean |
IMetaStoreClient.isPartitionMarkedForEvent(String db_name,
String tbl_name,
Map<String,String> partKVs,
PartitionEventType eventType) |
boolean |
HiveMetaStoreClient.isPartitionMarkedForEvent(String db_name,
String tbl_name,
Map<String,String> partKVs,
PartitionEventType eventType) |
boolean |
HiveMetaStore.HMSHandler.isPartitionMarkedForEvent(String db_name,
String tbl_name,
Map<String,String> partName,
PartitionEventType evtType) |
static boolean |
Deadline.isStarted()
Check whether the timer is started.
|
List<HiveObjectPrivilege> |
IMetaStoreClient.list_privileges(String principal_name,
PrincipalType principal_type,
HiveObjectRef hiveObject)
Return the privileges that this principal has directly over the object (not through roles).
|
List<HiveObjectPrivilege> |
HiveMetaStoreClient.list_privileges(String principalName,
PrincipalType principalType,
HiveObjectRef hiveObject) |
List<HiveObjectPrivilege> |
HiveMetaStore.HMSHandler.list_privileges(String principalName,
PrincipalType principalType,
HiveObjectRef hiveObject) |
List<Role> |
IMetaStoreClient.list_roles(String principalName,
PrincipalType principalType) |
List<Role> |
HiveMetaStoreClient.list_roles(String principalName,
PrincipalType principalType) |
List<Role> |
HiveMetaStore.HMSHandler.list_roles(String principalName,
PrincipalType principalType) |
List<Index> |
IMetaStoreClient.listIndexes(String db_name,
String tbl_name,
short max)
list indexes of the give base table
|
List<Index> |
HiveMetaStoreClient.listIndexes(String dbName,
String tblName,
short max)
list all the index names of the give base table.
|
List<String> |
RawStore.listIndexNames(String dbName,
String origTableName,
short max) |
List<String> |
ObjectStore.listIndexNames(String dbName,
String origTableName,
short max) |
List<String> |
IMetaStoreClient.listIndexNames(String db_name,
String tbl_name,
short max)
list all the index names of the give base table.
|
List<String> |
HiveMetaStoreClient.listIndexNames(String dbName,
String tblName,
short max)
list indexes of the give base table
|
List<String> |
IMetaStoreClient.listPartitionNames(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<String> |
HiveMetaStoreClient.listPartitionNames(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<String> |
RawStore.listPartitionNames(String db_name,
String tbl_name,
short max_parts) |
List<String> |
ObjectStore.listPartitionNames(String dbName,
String tableName,
short max) |
List<String> |
IMetaStoreClient.listPartitionNames(String db_name,
String tbl_name,
short max_parts) |
List<String> |
HiveMetaStoreClient.listPartitionNames(String dbName,
String tblName,
short max) |
List<String> |
RawStore.listPartitionNamesByFilter(String db_name,
String tbl_name,
String filter,
short max_parts) |
List<String> |
ObjectStore.listPartitionNamesByFilter(String dbName,
String tableName,
String filter,
short maxParts) |
List<String> |
RawStore.listPartitionNamesPs(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts)
Lists partition names that match a given partial specification
|
List<String> |
ObjectStore.listPartitionNamesPs(String dbName,
String tableName,
List<String> part_vals,
short max_parts) |
List<Partition> |
IMetaStoreClient.listPartitions(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<Partition> |
HiveMetaStoreClient.listPartitions(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<Partition> |
IMetaStoreClient.listPartitions(String db_name,
String tbl_name,
short max_parts) |
List<Partition> |
HiveMetaStoreClient.listPartitions(String db_name,
String tbl_name,
short max_parts) |
List<Partition> |
IMetaStoreClient.listPartitionsByFilter(String db_name,
String tbl_name,
String filter,
short max_parts)
Get list of partitions matching specified filter
|
List<Partition> |
HiveMetaStoreClient.listPartitionsByFilter(String db_name,
String tbl_name,
String filter,
short max_parts)
Get list of partitions matching specified filter
|
PartitionSpecProxy |
IMetaStoreClient.listPartitionSpecsByFilter(String db_name,
String tbl_name,
String filter,
int max_parts) |
PartitionSpecProxy |
HiveMetaStoreClient.listPartitionSpecsByFilter(String db_name,
String tbl_name,
String filter,
int max_parts) |
List<Partition> |
RawStore.listPartitionsPsWithAuth(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String userName,
List<String> groupNames)
Lists partitions that match a given partial specification and sets their auth privileges.
|
List<Partition> |
ObjectStore.listPartitionsPsWithAuth(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String userName,
List<String> groupNames) |
List<Partition> |
IMetaStoreClient.listPartitionsWithAuthInfo(String dbName,
String tableName,
List<String> partialPvals,
short s,
String userName,
List<String> groupNames) |
List<Partition> |
HiveMetaStoreClient.listPartitionsWithAuthInfo(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String user_name,
List<String> group_names) |
List<Partition> |
IMetaStoreClient.listPartitionsWithAuthInfo(String dbName,
String tableName,
short s,
String userName,
List<String> groupNames) |
List<Partition> |
HiveMetaStoreClient.listPartitionsWithAuthInfo(String db_name,
String tbl_name,
short max_parts,
String user_name,
List<String> group_names) |
List<String> |
IMetaStoreClient.listRoleNames()
list all role names
|
List<String> |
HiveMetaStoreClient.listRoleNames() |
List<String> |
RawStore.listTableNamesByFilter(String dbName,
String filter,
short max_tables)
Gets a list of tables based on a filter string and filter type.
|
List<String> |
ObjectStore.listTableNamesByFilter(String dbName,
String filter,
short maxTables) |
List<String> |
IMetaStoreClient.listTableNamesByFilter(String dbName,
String filter,
short maxTables)
Get a list of table names that match a filter.
|
List<String> |
HiveMetaStoreClient.listTableNamesByFilter(String dbName,
String filter,
short maxTables)
Get a list of table names that match a filter.
|
static void |
MetaStoreUtils.makeDir(org.apache.hadoop.fs.Path path,
HiveConf hiveConf) |
static Map<String,String> |
Warehouse.makeEscSpecFromName(String name) |
static String |
Warehouse.makePartName(List<FieldSchema> partCols,
List<String> vals) |
static String |
Warehouse.makePartName(List<FieldSchema> partCols,
List<String> vals,
String defaultStr)
Makes a valid partition name.
|
static String |
Warehouse.makePartName(Map<String,String> spec,
boolean addTrailingSeperator)
Makes a partition name from a specification
|
static String |
Warehouse.makePartPath(Map<String,String> spec)
Given a partition specification, return the path corresponding to the
partition spec.
|
static LinkedHashMap<String,String> |
Warehouse.makeSpecFromName(String name) |
static void |
Warehouse.makeValsFromName(String name,
AbstractList<String> result)
Extracts values from partition name without the column names.
|
Table |
RawStore.markPartitionForEvent(String dbName,
String tblName,
Map<String,String> partVals,
PartitionEventType evtType) |
Table |
ObjectStore.markPartitionForEvent(String dbName,
String tblName,
Map<String,String> partName,
PartitionEventType evtType) |
void |
IMetaStoreClient.markPartitionForEvent(String db_name,
String tbl_name,
Map<String,String> partKVs,
PartitionEventType eventType) |
void |
HiveMetaStoreClient.markPartitionForEvent(String db_name,
String tbl_name,
Map<String,String> partKVs,
PartitionEventType eventType) |
void |
HiveMetaStore.HMSHandler.markPartitionForEvent(String db_name,
String tbl_name,
Map<String,String> partName,
PartitionEventType evtType) |
boolean |
Warehouse.mkdirs(org.apache.hadoop.fs.Path f,
boolean inheritPermCandidate) |
static IHMSHandler |
HiveMetaStore.newRetryingHMSHandler(IHMSHandler baseHandler,
HiveConf hiveConf) |
static IHMSHandler |
HiveMetaStore.newRetryingHMSHandler(IHMSHandler baseHandler,
HiveConf hiveConf,
boolean local) |
static ThriftHiveMetastore.Iface |
HiveMetaStore.newRetryingHMSHandler(String name,
HiveConf conf,
boolean local) |
void |
MetaStoreEventListener.onAddIndex(AddIndexEvent indexEvent) |
void |
MetaStoreEventListener.onAddPartition(AddPartitionEvent partitionEvent) |
void |
MetaStoreEventListener.onAlterIndex(AlterIndexEvent indexEvent) |
void |
MetaStoreEventListener.onAlterPartition(AlterPartitionEvent partitionEvent) |
void |
MetaStoreEventListener.onAlterTable(AlterTableEvent tableEvent) |
void |
SessionPropertiesListener.onConfigChange(ConfigChangeEvent changeEvent) |
void |
MetaStoreEventListener.onConfigChange(ConfigChangeEvent tableEvent) |
void |
MetaStoreEventListener.onCreateDatabase(CreateDatabaseEvent dbEvent) |
void |
MetaStoreEventListener.onCreateTable(CreateTableEvent tableEvent) |
void |
MetaStoreEventListener.onDropDatabase(DropDatabaseEvent dbEvent) |
void |
MetaStoreEventListener.onDropIndex(DropIndexEvent indexEvent) |
void |
MetaStoreEventListener.onDropPartition(DropPartitionEvent partitionEvent) |
void |
MetaStoreEventListener.onDropTable(DropTableEvent tableEvent) |
abstract void |
MetaStorePreEventListener.onEvent(PreEventContext context) |
abstract void |
MetaStoreInitListener.onInit(MetaStoreInitContext context) |
void |
MetaStoreEventListener.onInsert(InsertEvent insertEvent)
This will be called when an insert is executed that does not cause a partition to be added.
|
void |
MetaStoreEventListener.onLoadPartitionDone(LoadPartitionDoneEvent partSetDoneEvent) |
boolean |
HiveMetaStore.HMSHandler.partition_name_has_valid_characters(List<String> part_vals,
boolean throw_exception) |
Map<String,String> |
HiveMetaStore.HMSHandler.partition_name_to_spec(String part_name) |
List<String> |
HiveMetaStore.HMSHandler.partition_name_to_vals(String part_name) |
Map<String,String> |
IMetaStoreClient.partitionNameToSpec(String name) |
Map<String,String> |
HiveMetaStoreClient.partitionNameToSpec(String name) |
List<String> |
IMetaStoreClient.partitionNameToVals(String name) |
List<String> |
HiveMetaStoreClient.partitionNameToVals(String name) |
void |
HiveMetaHook.preCreateTable(Table table)
Called before a new table definition is added to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.preDropTable(Table table)
Called before a table definition is removed from the metastore
during DROP TABLE.
|
void |
IMetaStoreClient.reconnect()
Tries to reconnect this MetaStoreClient to the MetaStore.
|
void |
HiveMetaStoreClient.reconnect() |
boolean |
RawStore.removeRole(String roleName) |
boolean |
ObjectStore.removeRole(String roleName) |
void |
HiveMetaStore.HMSHandler.rename_partition(String db_name,
String tbl_name,
List<String> part_vals,
Partition new_part) |
boolean |
Warehouse.renameDir(org.apache.hadoop.fs.Path sourcePath,
org.apache.hadoop.fs.Path destPath) |
boolean |
Warehouse.renameDir(org.apache.hadoop.fs.Path sourcePath,
org.apache.hadoop.fs.Path destPath,
boolean inheritPerms) |
void |
IMetaStoreClient.renamePartition(String dbname,
String name,
List<String> part_vals,
Partition newPart)
rename a partition to a new partition
|
void |
HiveMetaStoreClient.renamePartition(String dbname,
String name,
List<String> part_vals,
Partition newPart) |
long |
HiveMetaStore.HMSHandler.renew_delegation_token(String token_str_form) |
long |
IMetaStoreClient.renewDelegationToken(String tokenStrForm) |
long |
HiveMetaStoreClient.renewDelegationToken(String tokenStrForm) |
static void |
Deadline.resetTimeout(long timeout)
reset the timeout value of this timer.
|
boolean |
HiveMetaStore.HMSHandler.revoke_privileges(PrivilegeBag privileges) |
boolean |
IMetaStoreClient.revoke_privileges(PrivilegeBag privileges,
boolean grantOption) |
boolean |
HiveMetaStoreClient.revoke_privileges(PrivilegeBag privileges,
boolean grantOption) |
boolean |
HiveMetaStore.HMSHandler.revoke_privileges(PrivilegeBag privileges,
boolean grantOption) |
boolean |
HiveMetaStore.HMSHandler.revoke_role(String roleName,
String userName,
PrincipalType principalType) |
boolean |
IMetaStoreClient.revoke_role(String role_name,
String user_name,
PrincipalType principalType,
boolean grantOption) |
boolean |
HiveMetaStoreClient.revoke_role(String roleName,
String userName,
PrincipalType principalType,
boolean grantOption) |
boolean |
RawStore.revokePrivileges(PrivilegeBag privileges,
boolean grantOption) |
boolean |
ObjectStore.revokePrivileges(PrivilegeBag privileges,
boolean grantOption) |
boolean |
RawStore.revokeRole(Role role,
String userName,
PrincipalType principalType,
boolean grantOption) |
boolean |
ObjectStore.revokeRole(Role role,
String userName,
PrincipalType principalType,
boolean grantOption) |
void |
HiveMetaHook.rollbackCreateTable(Table table)
Called after failure adding a new table definition to the metastore
during CREATE TABLE.
|
void |
HiveMetaHook.rollbackDropTable(Table table)
Called after failure removing a table definition from the metastore
during DROP TABLE.
|
boolean |
HiveMetaStore.HMSHandler.set_aggr_stats_for(SetPartitionsStatsRequest request) |
List<String> |
HiveMetaStore.HMSHandler.set_ugi(String username,
List<String> groupNames) |
void |
IMetaStoreClient.setMetaConf(String key,
String value)
set meta variable which is open to end users
|
void |
HiveMetaStore.HMSHandler.setMetaConf(String key,
String value) |
void |
RawStore.setMetaStoreSchemaVersion(String version,
String comment) |
void |
ObjectStore.setMetaStoreSchemaVersion(String schemaVersion,
String comment) |
boolean |
IMetaStoreClient.setPartitionColumnStatistics(SetPartitionsStatsRequest request) |
boolean |
HiveMetaStoreClient.setPartitionColumnStatistics(SetPartitionsStatsRequest request) |
static void |
Deadline.startTimer(String method)
start the timer before a method is invoked.
|
static void |
Deadline.stopTimer()
end the time after a method is done.
|
boolean |
IMetaStoreClient.tableExists(String tableName)
Deprecated.
As of release 0.6.0 replaced by
IMetaStoreClient.tableExists(String, String) .
This method will be removed in release 0.7.0. |
boolean |
HiveMetaStoreClient.tableExists(String tableName)
Deprecated.
|
boolean |
IMetaStoreClient.tableExists(String databaseName,
String tableName) |
boolean |
HiveMetaStoreClient.tableExists(String databaseName,
String tableName) |
boolean |
HiveMetaStore.HMSHandler.update_partition_column_statistics(ColumnStatistics colStats) |
boolean |
HiveMetaStore.HMSHandler.update_table_column_statistics(ColumnStatistics colStats) |
void |
RawStore.updateMasterKey(Integer seqNo,
String key) |
void |
ObjectStore.updateMasterKey(Integer id,
String key) |
boolean |
IMetaStoreClient.updatePartitionColumnStatistics(ColumnStatistics statsObj)
Write partition level column statistics to persistent store
|
boolean |
HiveMetaStoreClient.updatePartitionColumnStatistics(ColumnStatistics statsObj)
Write partition level column statistics to persistent store
|
boolean |
RawStore.updatePartitionColumnStatistics(ColumnStatistics statsObj,
List<String> partVals)
Persists the given column statistics object to the metastore
|
boolean |
ObjectStore.updatePartitionColumnStatistics(ColumnStatistics colStats,
List<String> partVals) |
static boolean |
MetaStoreUtils.updatePartitionStatsFast(PartitionSpecProxy.PartitionIterator part,
Warehouse wh,
boolean madeDir,
boolean forceRecompute)
Updates the numFiles and totalSize parameters for the passed Partition by querying
the warehouse if the passed Partition does not already have values for these parameters.
|
static boolean |
MetaStoreUtils.updatePartitionStatsFast(Partition part,
Warehouse wh) |
static boolean |
MetaStoreUtils.updatePartitionStatsFast(Partition part,
Warehouse wh,
boolean madeDir) |
static boolean |
MetaStoreUtils.updatePartitionStatsFast(Partition part,
Warehouse wh,
boolean madeDir,
boolean forceRecompute)
Updates the numFiles and totalSize parameters for the passed Partition by querying
the warehouse if the passed Partition does not already have values for these parameters.
|
boolean |
RawStore.updateTableColumnStatistics(ColumnStatistics colStats)
Persists the given column statistics object to the metastore
|
boolean |
ObjectStore.updateTableColumnStatistics(ColumnStatistics colStats) |
boolean |
IMetaStoreClient.updateTableColumnStatistics(ColumnStatistics statsObj)
Write table level column statistics to persistent store
|
boolean |
HiveMetaStoreClient.updateTableColumnStatistics(ColumnStatistics statsObj)
Write table level column statistics to persistent store
|
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Database db,
Table tbl,
Warehouse wh,
boolean madeDir) |
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Database db,
Table tbl,
Warehouse wh,
boolean madeDir,
boolean forceRecompute) |
static boolean |
MetaStoreUtils.updateUnpartitionedTableStatsFast(Table tbl,
org.apache.hadoop.fs.FileStatus[] fileStatus,
boolean newDir,
boolean forceRecompute)
Updates the numFiles and totalSize parameters for the passed unpartitioned Table by querying
the warehouse if the passed Table does not already have values for these parameters.
|
void |
IMetaStoreClient.validatePartitionNameCharacters(List<String> partVals) |
void |
HiveMetaStoreClient.validatePartitionNameCharacters(List<String> partVals) |
static void |
MetaStoreUtils.validatePartitionNameCharacters(List<String> partVals,
Pattern partitionValidationPattern) |
void |
RawStore.verifySchema() |
void |
ObjectStore.verifySchema() |
Constructor and Description |
---|
HiveMetaStore.HMSHandler(String name) |
HiveMetaStore.HMSHandler(String name,
HiveConf conf) |
HiveMetaStore.HMSHandler(String name,
HiveConf conf,
boolean init) |
HiveMetaStoreClient(HiveConf conf) |
HiveMetaStoreClient(HiveConf conf,
HiveMetaHookLoader hookLoader) |
RawStoreProxy(HiveConf hiveConf,
org.apache.hadoop.conf.Configuration conf,
Class<? extends RawStore> rawStoreClass,
int id) |
RetryingMetaStoreClient(HiveConf hiveConf,
Class<?>[] constructorArgTypes,
Object[] constructorArgs,
Map<String,Long> metaCallTimeMap,
Class<? extends IMetaStoreClient> msClientClass) |
RetryingMetaStoreClient(HiveConf hiveConf,
HiveMetaHookLoader hookLoader,
Map<String,Long> metaCallTimeMap,
Class<? extends IMetaStoreClient> msClientClass) |
Warehouse(org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
MetaException |
MetaException.deepCopy() |
MetaException |
ThriftHiveMetastore.getMetaConf_result.getO1() |
MetaException |
ThriftHiveMetastore.setMetaConf_result.getO1() |
MetaException |
ThriftHiveMetastore.get_databases_result.getO1() |
MetaException |
ThriftHiveMetastore.get_all_databases_result.getO1() |
MetaException |
ThriftHiveMetastore.alter_database_result.getO1() |
MetaException |
ThriftHiveMetastore.get_type_result.getO1() |
MetaException |
ThriftHiveMetastore.drop_type_result.getO1() |
MetaException |
ThriftHiveMetastore.get_fields_result.getO1() |
MetaException |
ThriftHiveMetastore.get_fields_with_environment_context_result.getO1() |
MetaException |
ThriftHiveMetastore.get_schema_result.getO1() |
MetaException |
ThriftHiveMetastore.get_schema_with_environment_context_result.getO1() |
MetaException |
ThriftHiveMetastore.get_tables_result.getO1() |
MetaException |
ThriftHiveMetastore.get_all_tables_result.getO1() |
MetaException |
ThriftHiveMetastore.get_table_result.getO1() |
MetaException |
ThriftHiveMetastore.get_table_objects_by_name_result.getO1() |
MetaException |
ThriftHiveMetastore.get_table_names_by_filter_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partition_result.getO1() |
MetaException |
ThriftHiveMetastore.exchange_partition_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partition_with_auth_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partition_by_name_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partitions_ps_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partition_names_ps_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partitions_by_filter_result.getO1() |
MetaException |
ThriftHiveMetastore.get_part_specs_by_filter_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partitions_by_expr_result.getO1() |
MetaException |
ThriftHiveMetastore.get_partitions_by_names_result.getO1() |
MetaException |
ThriftHiveMetastore.partition_name_has_valid_characters_result.getO1() |
MetaException |
ThriftHiveMetastore.partition_name_to_vals_result.getO1() |
MetaException |
ThriftHiveMetastore.partition_name_to_spec_result.getO1() |
MetaException |
ThriftHiveMetastore.markPartitionForEvent_result.getO1() |
MetaException |
ThriftHiveMetastore.isPartitionMarkedForEvent_result.getO1() |
MetaException |
ThriftHiveMetastore.get_index_by_name_result.getO1() |
MetaException |
ThriftHiveMetastore.get_functions_result.getO1() |
MetaException |
ThriftHiveMetastore.get_function_result.getO1() |
MetaException |
ThriftHiveMetastore.create_role_result.getO1() |
MetaException |
ThriftHiveMetastore.drop_role_result.getO1() |
MetaException |
ThriftHiveMetastore.get_role_names_result.getO1() |
MetaException |
ThriftHiveMetastore.grant_role_result.getO1() |
MetaException |
ThriftHiveMetastore.revoke_role_result.getO1() |
MetaException |
ThriftHiveMetastore.list_roles_result.getO1() |
MetaException |
ThriftHiveMetastore.grant_revoke_role_result.getO1() |
MetaException |
ThriftHiveMetastore.get_principals_in_role_result.getO1() |
MetaException |
ThriftHiveMetastore.get_role_grants_for_principal_result.getO1() |
MetaException |
ThriftHiveMetastore.get_privilege_set_result.getO1() |
MetaException |
ThriftHiveMetastore.list_privileges_result.getO1() |
MetaException |
ThriftHiveMetastore.grant_privileges_result.getO1() |
MetaException |
ThriftHiveMetastore.revoke_privileges_result.getO1() |
MetaException |
ThriftHiveMetastore.grant_revoke_privileges_result.getO1() |
MetaException |
ThriftHiveMetastore.set_ugi_result.getO1() |
MetaException |
ThriftHiveMetastore.get_delegation_token_result.getO1() |
MetaException |
ThriftHiveMetastore.renew_delegation_token_result.getO1() |
MetaException |
ThriftHiveMetastore.cancel_delegation_token_result.getO1() |
MetaException |
ThriftHiveMetastore.get_database_result.getO2() |
MetaException |
ThriftHiveMetastore.get_type_all_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_table_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_table_with_environment_context_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_table_with_cascade_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_partition_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_partition_with_environment_context_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_partition_by_name_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_partition_by_name_with_environment_context_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_partitions_req_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partitions_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partitions_with_auth_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partitions_pspec_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partition_names_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partitions_ps_with_auth_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_partition_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_partitions_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_partition_with_environment_context_result.getO2() |
MetaException |
ThriftHiveMetastore.rename_partition_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_index_result.getO2() |
MetaException |
ThriftHiveMetastore.drop_index_by_name_result.getO2() |
MetaException |
ThriftHiveMetastore.get_indexes_result.getO2() |
MetaException |
ThriftHiveMetastore.get_index_names_result.getO2() |
MetaException |
ThriftHiveMetastore.get_table_column_statistics_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partition_column_statistics_result.getO2() |
MetaException |
ThriftHiveMetastore.get_table_statistics_req_result.getO2() |
MetaException |
ThriftHiveMetastore.get_partitions_statistics_req_result.getO2() |
MetaException |
ThriftHiveMetastore.get_aggr_stats_for_result.getO2() |
MetaException |
ThriftHiveMetastore.delete_partition_column_statistics_result.getO2() |
MetaException |
ThriftHiveMetastore.delete_table_column_statistics_result.getO2() |
MetaException |
ThriftHiveMetastore.alter_function_result.getO2() |
MetaException |
ThriftHiveMetastore.create_database_result.getO3() |
MetaException |
ThriftHiveMetastore.drop_database_result.getO3() |
MetaException |
ThriftHiveMetastore.create_type_result.getO3() |
MetaException |
ThriftHiveMetastore.create_table_result.getO3() |
MetaException |
ThriftHiveMetastore.create_table_with_environment_context_result.getO3() |
MetaException |
ThriftHiveMetastore.drop_table_result.getO3() |
MetaException |
ThriftHiveMetastore.drop_table_with_environment_context_result.getO3() |
MetaException |
ThriftHiveMetastore.add_partition_result.getO3() |
MetaException |
ThriftHiveMetastore.add_partition_with_environment_context_result.getO3() |
MetaException |
ThriftHiveMetastore.add_partitions_result.getO3() |
MetaException |
ThriftHiveMetastore.add_partitions_pspec_result.getO3() |
MetaException |
ThriftHiveMetastore.append_partition_result.getO3() |
MetaException |
ThriftHiveMetastore.add_partitions_req_result.getO3() |
MetaException |
ThriftHiveMetastore.append_partition_with_environment_context_result.getO3() |
MetaException |
ThriftHiveMetastore.append_partition_by_name_result.getO3() |
MetaException |
ThriftHiveMetastore.append_partition_by_name_with_environment_context_result.getO3() |
MetaException |
ThriftHiveMetastore.add_index_result.getO3() |
MetaException |
ThriftHiveMetastore.update_table_column_statistics_result.getO3() |
MetaException |
ThriftHiveMetastore.update_partition_column_statistics_result.getO3() |
MetaException |
ThriftHiveMetastore.set_aggr_stats_for_result.getO3() |
MetaException |
ThriftHiveMetastore.create_function_result.getO3() |
MetaException |
ThriftHiveMetastore.drop_function_result.getO3() |
Modifier and Type | Method and Description |
---|---|
int |
MetaException.compareTo(MetaException other) |
boolean |
MetaException.equals(MetaException that) |
void |
ThriftHiveMetastore.getMetaConf_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.setMetaConf_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_databases_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_all_databases_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.alter_database_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_type_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.drop_type_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_fields_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_fields_with_environment_context_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_schema_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_schema_with_environment_context_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_tables_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_all_tables_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_table_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_table_objects_by_name_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_table_names_by_filter_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partition_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.exchange_partition_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partition_with_auth_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partition_by_name_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partitions_ps_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partition_names_ps_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partitions_by_filter_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_part_specs_by_filter_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partitions_by_expr_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_partitions_by_names_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.partition_name_has_valid_characters_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.partition_name_to_vals_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.partition_name_to_spec_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.markPartitionForEvent_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.isPartitionMarkedForEvent_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_index_by_name_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_functions_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_function_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.create_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.drop_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_role_names_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.grant_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.revoke_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.list_roles_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.grant_revoke_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_principals_in_role_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_role_grants_for_principal_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_privilege_set_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.list_privileges_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.grant_privileges_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.revoke_privileges_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.grant_revoke_privileges_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.set_ugi_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_delegation_token_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.renew_delegation_token_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.cancel_delegation_token_result.setO1(MetaException o1) |
void |
ThriftHiveMetastore.get_database_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_type_all_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_table_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_table_with_environment_context_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_table_with_cascade_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_partition_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_partition_with_environment_context_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_partition_by_name_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_partition_by_name_with_environment_context_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_partitions_req_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partitions_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partitions_with_auth_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partitions_pspec_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partition_names_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partitions_ps_with_auth_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_partition_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_partitions_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_partition_with_environment_context_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.rename_partition_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_index_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.drop_index_by_name_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_indexes_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_index_names_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_table_column_statistics_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partition_column_statistics_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_table_statistics_req_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_partitions_statistics_req_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.get_aggr_stats_for_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.delete_partition_column_statistics_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.delete_table_column_statistics_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.alter_function_result.setO2(MetaException o2) |
void |
ThriftHiveMetastore.create_database_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.drop_database_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.create_type_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.create_table_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.create_table_with_environment_context_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.drop_table_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.drop_table_with_environment_context_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_partition_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_partition_with_environment_context_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_partitions_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_partitions_pspec_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.append_partition_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_partitions_req_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.append_partition_with_environment_context_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.append_partition_by_name_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.append_partition_by_name_with_environment_context_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.add_index_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.update_table_column_statistics_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.update_partition_column_statistics_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.set_aggr_stats_for_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.create_function_result.setO3(MetaException o3) |
void |
ThriftHiveMetastore.drop_function_result.setO3(MetaException o3) |
Modifier and Type | Method and Description |
---|---|
Index |
ThriftHiveMetastore.Iface.add_index(Index new_index,
Table index_table) |
Index |
ThriftHiveMetastore.Client.add_index(Index new_index,
Table index_table) |
Partition |
ThriftHiveMetastore.Iface.add_partition_with_environment_context(Partition new_part,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Client.add_partition_with_environment_context(Partition new_part,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Iface.add_partition(Partition new_part) |
Partition |
ThriftHiveMetastore.Client.add_partition(Partition new_part) |
int |
ThriftHiveMetastore.Iface.add_partitions_pspec(List<PartitionSpec> new_parts) |
int |
ThriftHiveMetastore.Client.add_partitions_pspec(List<PartitionSpec> new_parts) |
AddPartitionsResult |
ThriftHiveMetastore.Iface.add_partitions_req(AddPartitionsRequest request) |
AddPartitionsResult |
ThriftHiveMetastore.Client.add_partitions_req(AddPartitionsRequest request) |
int |
ThriftHiveMetastore.Iface.add_partitions(List<Partition> new_parts) |
int |
ThriftHiveMetastore.Client.add_partitions(List<Partition> new_parts) |
void |
ThriftHiveMetastore.Iface.alter_database(String dbname,
Database db) |
void |
ThriftHiveMetastore.Client.alter_database(String dbname,
Database db) |
void |
ThriftHiveMetastore.Iface.alter_function(String dbName,
String funcName,
Function newFunc) |
void |
ThriftHiveMetastore.Client.alter_function(String dbName,
String funcName,
Function newFunc) |
void |
ThriftHiveMetastore.Iface.alter_index(String dbname,
String base_tbl_name,
String idx_name,
Index new_idx) |
void |
ThriftHiveMetastore.Client.alter_index(String dbname,
String base_tbl_name,
String idx_name,
Index new_idx) |
void |
ThriftHiveMetastore.Iface.alter_partition_with_environment_context(String db_name,
String tbl_name,
Partition new_part,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.alter_partition_with_environment_context(String db_name,
String tbl_name,
Partition new_part,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Iface.alter_partition(String db_name,
String tbl_name,
Partition new_part) |
void |
ThriftHiveMetastore.Client.alter_partition(String db_name,
String tbl_name,
Partition new_part) |
void |
ThriftHiveMetastore.Iface.alter_partitions(String db_name,
String tbl_name,
List<Partition> new_parts) |
void |
ThriftHiveMetastore.Client.alter_partitions(String db_name,
String tbl_name,
List<Partition> new_parts) |
void |
ThriftHiveMetastore.Iface.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
ThriftHiveMetastore.Client.alter_table_with_cascade(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
ThriftHiveMetastore.Iface.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.alter_table_with_environment_context(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Iface.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
void |
ThriftHiveMetastore.Client.alter_table(String dbname,
String tbl_name,
Table new_tbl) |
Partition |
ThriftHiveMetastore.Iface.append_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Client.append_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Iface.append_partition_by_name(String db_name,
String tbl_name,
String part_name) |
Partition |
ThriftHiveMetastore.Client.append_partition_by_name(String db_name,
String tbl_name,
String part_name) |
Partition |
ThriftHiveMetastore.Iface.append_partition_with_environment_context(String db_name,
String tbl_name,
List<String> part_vals,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Client.append_partition_with_environment_context(String db_name,
String tbl_name,
List<String> part_vals,
EnvironmentContext environment_context) |
Partition |
ThriftHiveMetastore.Iface.append_partition(String db_name,
String tbl_name,
List<String> part_vals) |
Partition |
ThriftHiveMetastore.Client.append_partition(String db_name,
String tbl_name,
List<String> part_vals) |
void |
ThriftHiveMetastore.Iface.cancel_delegation_token(String token_str_form) |
void |
ThriftHiveMetastore.Client.cancel_delegation_token(String token_str_form) |
void |
ThriftHiveMetastore.Iface.create_database(Database database) |
void |
ThriftHiveMetastore.Client.create_database(Database database) |
void |
ThriftHiveMetastore.Iface.create_function(Function func) |
void |
ThriftHiveMetastore.Client.create_function(Function func) |
boolean |
ThriftHiveMetastore.Iface.create_role(Role role) |
boolean |
ThriftHiveMetastore.Client.create_role(Role role) |
void |
ThriftHiveMetastore.Iface.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.create_table_with_environment_context(Table tbl,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Iface.create_table(Table tbl) |
void |
ThriftHiveMetastore.Client.create_table(Table tbl) |
boolean |
ThriftHiveMetastore.Iface.create_type(Type type) |
boolean |
ThriftHiveMetastore.Client.create_type(Type type) |
boolean |
ThriftHiveMetastore.Iface.delete_partition_column_statistics(String db_name,
String tbl_name,
String part_name,
String col_name) |
boolean |
ThriftHiveMetastore.Client.delete_partition_column_statistics(String db_name,
String tbl_name,
String part_name,
String col_name) |
boolean |
ThriftHiveMetastore.Iface.delete_table_column_statistics(String db_name,
String tbl_name,
String col_name) |
boolean |
ThriftHiveMetastore.Client.delete_table_column_statistics(String db_name,
String tbl_name,
String col_name) |
void |
ThriftHiveMetastore.Iface.drop_database(String name,
boolean deleteData,
boolean cascade) |
void |
ThriftHiveMetastore.Client.drop_database(String name,
boolean deleteData,
boolean cascade) |
void |
ThriftHiveMetastore.Iface.drop_function(String dbName,
String funcName) |
void |
ThriftHiveMetastore.Client.drop_function(String dbName,
String funcName) |
boolean |
ThriftHiveMetastore.Iface.drop_index_by_name(String db_name,
String tbl_name,
String index_name,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Client.drop_index_by_name(String db_name,
String tbl_name,
String index_name,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Iface.drop_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
boolean deleteData,
EnvironmentContext environment_context) |
boolean |
ThriftHiveMetastore.Client.drop_partition_by_name_with_environment_context(String db_name,
String tbl_name,
String part_name,
boolean deleteData,
EnvironmentContext environment_context) |
boolean |
ThriftHiveMetastore.Iface.drop_partition_by_name(String db_name,
String tbl_name,
String part_name,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Client.drop_partition_by_name(String db_name,
String tbl_name,
String part_name,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Iface.drop_partition_with_environment_context(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData,
EnvironmentContext environment_context) |
boolean |
ThriftHiveMetastore.Client.drop_partition_with_environment_context(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData,
EnvironmentContext environment_context) |
boolean |
ThriftHiveMetastore.Iface.drop_partition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Client.drop_partition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData) |
DropPartitionsResult |
ThriftHiveMetastore.Iface.drop_partitions_req(DropPartitionsRequest req) |
DropPartitionsResult |
ThriftHiveMetastore.Client.drop_partitions_req(DropPartitionsRequest req) |
boolean |
ThriftHiveMetastore.Iface.drop_role(String role_name) |
boolean |
ThriftHiveMetastore.Client.drop_role(String role_name) |
void |
ThriftHiveMetastore.Iface.drop_table_with_environment_context(String dbname,
String name,
boolean deleteData,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Client.drop_table_with_environment_context(String dbname,
String name,
boolean deleteData,
EnvironmentContext environment_context) |
void |
ThriftHiveMetastore.Iface.drop_table(String dbname,
String name,
boolean deleteData) |
void |
ThriftHiveMetastore.Client.drop_table(String dbname,
String name,
boolean deleteData) |
boolean |
ThriftHiveMetastore.Iface.drop_type(String type) |
boolean |
ThriftHiveMetastore.Client.drop_type(String type) |
Partition |
ThriftHiveMetastore.Iface.exchange_partition(Map<String,String> partitionSpecs,
String source_db,
String source_table_name,
String dest_db,
String dest_table_name) |
Partition |
ThriftHiveMetastore.Client.exchange_partition(Map<String,String> partitionSpecs,
String source_db,
String source_table_name,
String dest_db,
String dest_table_name) |
AggrStats |
ThriftHiveMetastore.Iface.get_aggr_stats_for(PartitionsStatsRequest request) |
AggrStats |
ThriftHiveMetastore.Client.get_aggr_stats_for(PartitionsStatsRequest request) |
List<String> |
ThriftHiveMetastore.Iface.get_all_databases() |
List<String> |
ThriftHiveMetastore.Client.get_all_databases() |
List<String> |
ThriftHiveMetastore.Iface.get_all_tables(String db_name) |
List<String> |
ThriftHiveMetastore.Client.get_all_tables(String db_name) |
Database |
ThriftHiveMetastore.Iface.get_database(String name) |
Database |
ThriftHiveMetastore.Client.get_database(String name) |
List<String> |
ThriftHiveMetastore.Iface.get_databases(String pattern) |
List<String> |
ThriftHiveMetastore.Client.get_databases(String pattern) |
String |
ThriftHiveMetastore.Iface.get_delegation_token(String token_owner,
String renewer_kerberos_principal_name) |
String |
ThriftHiveMetastore.Client.get_delegation_token(String token_owner,
String renewer_kerberos_principal_name) |
List<FieldSchema> |
ThriftHiveMetastore.Iface.get_fields_with_environment_context(String db_name,
String table_name,
EnvironmentContext environment_context) |
List<FieldSchema> |
ThriftHiveMetastore.Client.get_fields_with_environment_context(String db_name,
String table_name,
EnvironmentContext environment_context) |
List<FieldSchema> |
ThriftHiveMetastore.Iface.get_fields(String db_name,
String table_name) |
List<FieldSchema> |
ThriftHiveMetastore.Client.get_fields(String db_name,
String table_name) |
Function |
ThriftHiveMetastore.Iface.get_function(String dbName,
String funcName) |
Function |
ThriftHiveMetastore.Client.get_function(String dbName,
String funcName) |
List<String> |
ThriftHiveMetastore.Iface.get_functions(String dbName,
String pattern) |
List<String> |
ThriftHiveMetastore.Client.get_functions(String dbName,
String pattern) |
Index |
ThriftHiveMetastore.Iface.get_index_by_name(String db_name,
String tbl_name,
String index_name) |
Index |
ThriftHiveMetastore.Client.get_index_by_name(String db_name,
String tbl_name,
String index_name) |
List<String> |
ThriftHiveMetastore.Iface.get_index_names(String db_name,
String tbl_name,
short max_indexes) |
List<String> |
ThriftHiveMetastore.Client.get_index_names(String db_name,
String tbl_name,
short max_indexes) |
List<Index> |
ThriftHiveMetastore.Iface.get_indexes(String db_name,
String tbl_name,
short max_indexes) |
List<Index> |
ThriftHiveMetastore.Client.get_indexes(String db_name,
String tbl_name,
short max_indexes) |
List<PartitionSpec> |
ThriftHiveMetastore.Iface.get_part_specs_by_filter(String db_name,
String tbl_name,
String filter,
int max_parts) |
List<PartitionSpec> |
ThriftHiveMetastore.Client.get_part_specs_by_filter(String db_name,
String tbl_name,
String filter,
int max_parts) |
Partition |
ThriftHiveMetastore.Iface.get_partition_by_name(String db_name,
String tbl_name,
String part_name) |
Partition |
ThriftHiveMetastore.Client.get_partition_by_name(String db_name,
String tbl_name,
String part_name) |
ColumnStatistics |
ThriftHiveMetastore.Iface.get_partition_column_statistics(String db_name,
String tbl_name,
String part_name,
String col_name) |
ColumnStatistics |
ThriftHiveMetastore.Client.get_partition_column_statistics(String db_name,
String tbl_name,
String part_name,
String col_name) |
List<String> |
ThriftHiveMetastore.Iface.get_partition_names_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<String> |
ThriftHiveMetastore.Client.get_partition_names_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<String> |
ThriftHiveMetastore.Iface.get_partition_names(String db_name,
String tbl_name,
short max_parts) |
List<String> |
ThriftHiveMetastore.Client.get_partition_names(String db_name,
String tbl_name,
short max_parts) |
Partition |
ThriftHiveMetastore.Iface.get_partition_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
String user_name,
List<String> group_names) |
Partition |
ThriftHiveMetastore.Client.get_partition_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
String user_name,
List<String> group_names) |
Partition |
ThriftHiveMetastore.Iface.get_partition(String db_name,
String tbl_name,
List<String> part_vals) |
Partition |
ThriftHiveMetastore.Client.get_partition(String db_name,
String tbl_name,
List<String> part_vals) |
PartitionsByExprResult |
ThriftHiveMetastore.Iface.get_partitions_by_expr(PartitionsByExprRequest req) |
PartitionsByExprResult |
ThriftHiveMetastore.Client.get_partitions_by_expr(PartitionsByExprRequest req) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions_by_filter(String db_name,
String tbl_name,
String filter,
short max_parts) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions_by_filter(String db_name,
String tbl_name,
String filter,
short max_parts) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions_by_names(String db_name,
String tbl_name,
List<String> names) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions_by_names(String db_name,
String tbl_name,
List<String> names) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions_ps_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String user_name,
List<String> group_names) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions_ps_with_auth(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts,
String user_name,
List<String> group_names) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions_ps(String db_name,
String tbl_name,
List<String> part_vals,
short max_parts) |
List<PartitionSpec> |
ThriftHiveMetastore.Iface.get_partitions_pspec(String db_name,
String tbl_name,
int max_parts) |
List<PartitionSpec> |
ThriftHiveMetastore.Client.get_partitions_pspec(String db_name,
String tbl_name,
int max_parts) |
PartitionsStatsResult |
ThriftHiveMetastore.Iface.get_partitions_statistics_req(PartitionsStatsRequest request) |
PartitionsStatsResult |
ThriftHiveMetastore.Client.get_partitions_statistics_req(PartitionsStatsRequest request) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions_with_auth(String db_name,
String tbl_name,
short max_parts,
String user_name,
List<String> group_names) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions_with_auth(String db_name,
String tbl_name,
short max_parts,
String user_name,
List<String> group_names) |
List<Partition> |
ThriftHiveMetastore.Iface.get_partitions(String db_name,
String tbl_name,
short max_parts) |
List<Partition> |
ThriftHiveMetastore.Client.get_partitions(String db_name,
String tbl_name,
short max_parts) |
GetPrincipalsInRoleResponse |
ThriftHiveMetastore.Iface.get_principals_in_role(GetPrincipalsInRoleRequest request) |
GetPrincipalsInRoleResponse |
ThriftHiveMetastore.Client.get_principals_in_role(GetPrincipalsInRoleRequest request) |
PrincipalPrivilegeSet |
ThriftHiveMetastore.Iface.get_privilege_set(HiveObjectRef hiveObject,
String user_name,
List<String> group_names) |
PrincipalPrivilegeSet |
ThriftHiveMetastore.Client.get_privilege_set(HiveObjectRef hiveObject,
String user_name,
List<String> group_names) |
GetRoleGrantsForPrincipalResponse |
ThriftHiveMetastore.Iface.get_role_grants_for_principal(GetRoleGrantsForPrincipalRequest request) |
GetRoleGrantsForPrincipalResponse |
ThriftHiveMetastore.Client.get_role_grants_for_principal(GetRoleGrantsForPrincipalRequest request) |
List<String> |
ThriftHiveMetastore.Iface.get_role_names() |
List<String> |
ThriftHiveMetastore.Client.get_role_names() |
List<FieldSchema> |
ThriftHiveMetastore.Iface.get_schema_with_environment_context(String db_name,
String table_name,
EnvironmentContext environment_context) |
List<FieldSchema> |
ThriftHiveMetastore.Client.get_schema_with_environment_context(String db_name,
String table_name,
EnvironmentContext environment_context) |
List<FieldSchema> |
ThriftHiveMetastore.Iface.get_schema(String db_name,
String table_name) |
List<FieldSchema> |
ThriftHiveMetastore.Client.get_schema(String db_name,
String table_name) |
ColumnStatistics |
ThriftHiveMetastore.Iface.get_table_column_statistics(String db_name,
String tbl_name,
String col_name) |
ColumnStatistics |
ThriftHiveMetastore.Client.get_table_column_statistics(String db_name,
String tbl_name,
String col_name) |
List<String> |
ThriftHiveMetastore.Iface.get_table_names_by_filter(String dbname,
String filter,
short max_tables) |
List<String> |
ThriftHiveMetastore.Client.get_table_names_by_filter(String dbname,
String filter,
short max_tables) |
List<Table> |
ThriftHiveMetastore.Iface.get_table_objects_by_name(String dbname,
List<String> tbl_names) |
List<Table> |
ThriftHiveMetastore.Client.get_table_objects_by_name(String dbname,
List<String> tbl_names) |
TableStatsResult |
ThriftHiveMetastore.Iface.get_table_statistics_req(TableStatsRequest request) |
TableStatsResult |
ThriftHiveMetastore.Client.get_table_statistics_req(TableStatsRequest request) |
Table |
ThriftHiveMetastore.Iface.get_table(String dbname,
String tbl_name) |
Table |
ThriftHiveMetastore.Client.get_table(String dbname,
String tbl_name) |
List<String> |
ThriftHiveMetastore.Iface.get_tables(String db_name,
String pattern) |
List<String> |
ThriftHiveMetastore.Client.get_tables(String db_name,
String pattern) |
Map<String,Type> |
ThriftHiveMetastore.Iface.get_type_all(String name) |
Map<String,Type> |
ThriftHiveMetastore.Client.get_type_all(String name) |
Type |
ThriftHiveMetastore.Iface.get_type(String name) |
Type |
ThriftHiveMetastore.Client.get_type(String name) |
String |
ThriftHiveMetastore.Iface.getMetaConf(String key) |
String |
ThriftHiveMetastore.Client.getMetaConf(String key) |
String |
ThriftHiveMetastore.AsyncClient.getMetaConf_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.setMetaConf_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.create_database_call.getResult() |
Database |
ThriftHiveMetastore.AsyncClient.get_database_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.drop_database_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_databases_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_all_databases_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_database_call.getResult() |
Type |
ThriftHiveMetastore.AsyncClient.get_type_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.create_type_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_type_call.getResult() |
Map<String,Type> |
ThriftHiveMetastore.AsyncClient.get_type_all_call.getResult() |
List<FieldSchema> |
ThriftHiveMetastore.AsyncClient.get_fields_call.getResult() |
List<FieldSchema> |
ThriftHiveMetastore.AsyncClient.get_fields_with_environment_context_call.getResult() |
List<FieldSchema> |
ThriftHiveMetastore.AsyncClient.get_schema_call.getResult() |
List<FieldSchema> |
ThriftHiveMetastore.AsyncClient.get_schema_with_environment_context_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.create_table_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.create_table_with_environment_context_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.drop_table_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.drop_table_with_environment_context_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_tables_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_all_tables_call.getResult() |
Table |
ThriftHiveMetastore.AsyncClient.get_table_call.getResult() |
List<Table> |
ThriftHiveMetastore.AsyncClient.get_table_objects_by_name_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_table_names_by_filter_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_table_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_table_with_environment_context_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_table_with_cascade_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.add_partition_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.add_partition_with_environment_context_call.getResult() |
int |
ThriftHiveMetastore.AsyncClient.add_partitions_call.getResult() |
int |
ThriftHiveMetastore.AsyncClient.add_partitions_pspec_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.append_partition_call.getResult() |
AddPartitionsResult |
ThriftHiveMetastore.AsyncClient.add_partitions_req_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.append_partition_with_environment_context_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.append_partition_by_name_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.append_partition_by_name_with_environment_context_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_partition_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_partition_with_environment_context_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_partition_by_name_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_partition_by_name_with_environment_context_call.getResult() |
DropPartitionsResult |
ThriftHiveMetastore.AsyncClient.drop_partitions_req_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.get_partition_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.exchange_partition_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.get_partition_with_auth_call.getResult() |
Partition |
ThriftHiveMetastore.AsyncClient.get_partition_by_name_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_with_auth_call.getResult() |
List<PartitionSpec> |
ThriftHiveMetastore.AsyncClient.get_partitions_pspec_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_partition_names_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_ps_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_ps_with_auth_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_partition_names_ps_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_by_filter_call.getResult() |
List<PartitionSpec> |
ThriftHiveMetastore.AsyncClient.get_part_specs_by_filter_call.getResult() |
PartitionsByExprResult |
ThriftHiveMetastore.AsyncClient.get_partitions_by_expr_call.getResult() |
List<Partition> |
ThriftHiveMetastore.AsyncClient.get_partitions_by_names_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_partition_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_partitions_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_partition_with_environment_context_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.rename_partition_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.partition_name_has_valid_characters_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.partition_name_to_vals_call.getResult() |
Map<String,String> |
ThriftHiveMetastore.AsyncClient.partition_name_to_spec_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.markPartitionForEvent_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.isPartitionMarkedForEvent_call.getResult() |
Index |
ThriftHiveMetastore.AsyncClient.add_index_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_index_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_index_by_name_call.getResult() |
Index |
ThriftHiveMetastore.AsyncClient.get_index_by_name_call.getResult() |
List<Index> |
ThriftHiveMetastore.AsyncClient.get_indexes_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_index_names_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.update_table_column_statistics_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.update_partition_column_statistics_call.getResult() |
ColumnStatistics |
ThriftHiveMetastore.AsyncClient.get_table_column_statistics_call.getResult() |
ColumnStatistics |
ThriftHiveMetastore.AsyncClient.get_partition_column_statistics_call.getResult() |
TableStatsResult |
ThriftHiveMetastore.AsyncClient.get_table_statistics_req_call.getResult() |
PartitionsStatsResult |
ThriftHiveMetastore.AsyncClient.get_partitions_statistics_req_call.getResult() |
AggrStats |
ThriftHiveMetastore.AsyncClient.get_aggr_stats_for_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.set_aggr_stats_for_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.delete_partition_column_statistics_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.delete_table_column_statistics_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.create_function_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.drop_function_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.alter_function_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_functions_call.getResult() |
Function |
ThriftHiveMetastore.AsyncClient.get_function_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.create_role_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.drop_role_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.get_role_names_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.grant_role_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.revoke_role_call.getResult() |
List<Role> |
ThriftHiveMetastore.AsyncClient.list_roles_call.getResult() |
GrantRevokeRoleResponse |
ThriftHiveMetastore.AsyncClient.grant_revoke_role_call.getResult() |
GetPrincipalsInRoleResponse |
ThriftHiveMetastore.AsyncClient.get_principals_in_role_call.getResult() |
GetRoleGrantsForPrincipalResponse |
ThriftHiveMetastore.AsyncClient.get_role_grants_for_principal_call.getResult() |
PrincipalPrivilegeSet |
ThriftHiveMetastore.AsyncClient.get_privilege_set_call.getResult() |
List<HiveObjectPrivilege> |
ThriftHiveMetastore.AsyncClient.list_privileges_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.grant_privileges_call.getResult() |
boolean |
ThriftHiveMetastore.AsyncClient.revoke_privileges_call.getResult() |
GrantRevokePrivilegeResponse |
ThriftHiveMetastore.AsyncClient.grant_revoke_privileges_call.getResult() |
List<String> |
ThriftHiveMetastore.AsyncClient.set_ugi_call.getResult() |
String |
ThriftHiveMetastore.AsyncClient.get_delegation_token_call.getResult() |
long |
ThriftHiveMetastore.AsyncClient.renew_delegation_token_call.getResult() |
void |
ThriftHiveMetastore.AsyncClient.cancel_delegation_token_call.getResult() |
boolean |
ThriftHiveMetastore.Iface.grant_privileges(PrivilegeBag privileges) |
boolean |
ThriftHiveMetastore.Client.grant_privileges(PrivilegeBag privileges) |
GrantRevokePrivilegeResponse |
ThriftHiveMetastore.Iface.grant_revoke_privileges(GrantRevokePrivilegeRequest request) |
GrantRevokePrivilegeResponse |
ThriftHiveMetastore.Client.grant_revoke_privileges(GrantRevokePrivilegeRequest request) |
GrantRevokeRoleResponse |
ThriftHiveMetastore.Iface.grant_revoke_role(GrantRevokeRoleRequest request) |
GrantRevokeRoleResponse |
ThriftHiveMetastore.Client.grant_revoke_role(GrantRevokeRoleRequest request) |
boolean |
ThriftHiveMetastore.Iface.grant_role(String role_name,
String principal_name,
PrincipalType principal_type,
String grantor,
PrincipalType grantorType,
boolean grant_option) |
boolean |
ThriftHiveMetastore.Client.grant_role(String role_name,
String principal_name,
PrincipalType principal_type,
String grantor,
PrincipalType grantorType,
boolean grant_option) |
boolean |
ThriftHiveMetastore.Iface.isPartitionMarkedForEvent(String db_name,
String tbl_name,
Map<String,String> part_vals,
PartitionEventType eventType) |
boolean |
ThriftHiveMetastore.Client.isPartitionMarkedForEvent(String db_name,
String tbl_name,
Map<String,String> part_vals,
PartitionEventType eventType) |
List<HiveObjectPrivilege> |
ThriftHiveMetastore.Iface.list_privileges(String principal_name,
PrincipalType principal_type,
HiveObjectRef hiveObject) |
List<HiveObjectPrivilege> |
ThriftHiveMetastore.Client.list_privileges(String principal_name,
PrincipalType principal_type,
HiveObjectRef hiveObject) |
List<Role> |
ThriftHiveMetastore.Iface.list_roles(String principal_name,
PrincipalType principal_type) |
List<Role> |
ThriftHiveMetastore.Client.list_roles(String principal_name,
PrincipalType principal_type) |
void |
ThriftHiveMetastore.Iface.markPartitionForEvent(String db_name,
String tbl_name,
Map<String,String> part_vals,
PartitionEventType eventType) |
void |
ThriftHiveMetastore.Client.markPartitionForEvent(String db_name,
String tbl_name,
Map<String,String> part_vals,
PartitionEventType eventType) |
boolean |
ThriftHiveMetastore.Iface.partition_name_has_valid_characters(List<String> part_vals,
boolean throw_exception) |
boolean |
ThriftHiveMetastore.Client.partition_name_has_valid_characters(List<String> part_vals,
boolean throw_exception) |
Map<String,String> |
ThriftHiveMetastore.Iface.partition_name_to_spec(String part_name) |
Map<String,String> |
ThriftHiveMetastore.Client.partition_name_to_spec(String part_name) |
List<String> |
ThriftHiveMetastore.Iface.partition_name_to_vals(String part_name) |
List<String> |
ThriftHiveMetastore.Client.partition_name_to_vals(String part_name) |
Index |
ThriftHiveMetastore.Client.recv_add_index() |
Partition |
ThriftHiveMetastore.Client.recv_add_partition_with_environment_context() |
Partition |
ThriftHiveMetastore.Client.recv_add_partition() |
int |
ThriftHiveMetastore.Client.recv_add_partitions_pspec() |
AddPartitionsResult |
ThriftHiveMetastore.Client.recv_add_partitions_req() |
int |
ThriftHiveMetastore.Client.recv_add_partitions() |
void |
ThriftHiveMetastore.Client.recv_alter_database() |
void |
ThriftHiveMetastore.Client.recv_alter_function() |
void |
ThriftHiveMetastore.Client.recv_alter_index() |
void |
ThriftHiveMetastore.Client.recv_alter_partition_with_environment_context() |
void |
ThriftHiveMetastore.Client.recv_alter_partition() |
void |
ThriftHiveMetastore.Client.recv_alter_partitions() |
void |
ThriftHiveMetastore.Client.recv_alter_table_with_cascade() |
void |
ThriftHiveMetastore.Client.recv_alter_table_with_environment_context() |
void |
ThriftHiveMetastore.Client.recv_alter_table() |
Partition |
ThriftHiveMetastore.Client.recv_append_partition_by_name_with_environment_context() |
Partition |
ThriftHiveMetastore.Client.recv_append_partition_by_name() |
Partition |
ThriftHiveMetastore.Client.recv_append_partition_with_environment_context() |
Partition |
ThriftHiveMetastore.Client.recv_append_partition() |
void |
ThriftHiveMetastore.Client.recv_cancel_delegation_token() |
void |
ThriftHiveMetastore.Client.recv_create_database() |
void |
ThriftHiveMetastore.Client.recv_create_function() |
boolean |
ThriftHiveMetastore.Client.recv_create_role() |
void |
ThriftHiveMetastore.Client.recv_create_table_with_environment_context() |
void |
ThriftHiveMetastore.Client.recv_create_table() |
boolean |
ThriftHiveMetastore.Client.recv_create_type() |
boolean |
ThriftHiveMetastore.Client.recv_delete_partition_column_statistics() |
boolean |
ThriftHiveMetastore.Client.recv_delete_table_column_statistics() |
void |
ThriftHiveMetastore.Client.recv_drop_database() |
void |
ThriftHiveMetastore.Client.recv_drop_function() |
boolean |
ThriftHiveMetastore.Client.recv_drop_index_by_name() |
boolean |
ThriftHiveMetastore.Client.recv_drop_partition_by_name_with_environment_context() |
boolean |
ThriftHiveMetastore.Client.recv_drop_partition_by_name() |
boolean |
ThriftHiveMetastore.Client.recv_drop_partition_with_environment_context() |
boolean |
ThriftHiveMetastore.Client.recv_drop_partition() |
DropPartitionsResult |
ThriftHiveMetastore.Client.recv_drop_partitions_req() |
boolean |
ThriftHiveMetastore.Client.recv_drop_role() |
void |
ThriftHiveMetastore.Client.recv_drop_table_with_environment_context() |
void |
ThriftHiveMetastore.Client.recv_drop_table() |
boolean |
ThriftHiveMetastore.Client.recv_drop_type() |
Partition |
ThriftHiveMetastore.Client.recv_exchange_partition() |
AggrStats |
ThriftHiveMetastore.Client.recv_get_aggr_stats_for() |
List<String> |
ThriftHiveMetastore.Client.recv_get_all_databases() |
List<String> |
ThriftHiveMetastore.Client.recv_get_all_tables() |
Database |
ThriftHiveMetastore.Client.recv_get_database() |
List<String> |
ThriftHiveMetastore.Client.recv_get_databases() |
String |
ThriftHiveMetastore.Client.recv_get_delegation_token() |
List<FieldSchema> |
ThriftHiveMetastore.Client.recv_get_fields_with_environment_context() |
List<FieldSchema> |
ThriftHiveMetastore.Client.recv_get_fields() |
Function |
ThriftHiveMetastore.Client.recv_get_function() |
List<String> |
ThriftHiveMetastore.Client.recv_get_functions() |
Index |
ThriftHiveMetastore.Client.recv_get_index_by_name() |
List<String> |
ThriftHiveMetastore.Client.recv_get_index_names() |
List<Index> |
ThriftHiveMetastore.Client.recv_get_indexes() |
List<PartitionSpec> |
ThriftHiveMetastore.Client.recv_get_part_specs_by_filter() |
Partition |
ThriftHiveMetastore.Client.recv_get_partition_by_name() |
ColumnStatistics |
ThriftHiveMetastore.Client.recv_get_partition_column_statistics() |
List<String> |
ThriftHiveMetastore.Client.recv_get_partition_names_ps() |
List<String> |
ThriftHiveMetastore.Client.recv_get_partition_names() |
Partition |
ThriftHiveMetastore.Client.recv_get_partition_with_auth() |
Partition |
ThriftHiveMetastore.Client.recv_get_partition() |
PartitionsByExprResult |
ThriftHiveMetastore.Client.recv_get_partitions_by_expr() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions_by_filter() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions_by_names() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions_ps_with_auth() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions_ps() |
List<PartitionSpec> |
ThriftHiveMetastore.Client.recv_get_partitions_pspec() |
PartitionsStatsResult |
ThriftHiveMetastore.Client.recv_get_partitions_statistics_req() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions_with_auth() |
List<Partition> |
ThriftHiveMetastore.Client.recv_get_partitions() |
GetPrincipalsInRoleResponse |
ThriftHiveMetastore.Client.recv_get_principals_in_role() |
PrincipalPrivilegeSet |
ThriftHiveMetastore.Client.recv_get_privilege_set() |
GetRoleGrantsForPrincipalResponse |
ThriftHiveMetastore.Client.recv_get_role_grants_for_principal() |
List<String> |
ThriftHiveMetastore.Client.recv_get_role_names() |
List<FieldSchema> |
ThriftHiveMetastore.Client.recv_get_schema_with_environment_context() |
List<FieldSchema> |
ThriftHiveMetastore.Client.recv_get_schema() |
ColumnStatistics |
ThriftHiveMetastore.Client.recv_get_table_column_statistics() |
List<String> |
ThriftHiveMetastore.Client.recv_get_table_names_by_filter() |
List<Table> |
ThriftHiveMetastore.Client.recv_get_table_objects_by_name() |
TableStatsResult |
ThriftHiveMetastore.Client.recv_get_table_statistics_req() |
Table |
ThriftHiveMetastore.Client.recv_get_table() |
List<String> |
ThriftHiveMetastore.Client.recv_get_tables() |
Map<String,Type> |
ThriftHiveMetastore.Client.recv_get_type_all() |
Type |
ThriftHiveMetastore.Client.recv_get_type() |
String |
ThriftHiveMetastore.Client.recv_getMetaConf() |
boolean |
ThriftHiveMetastore.Client.recv_grant_privileges() |
GrantRevokePrivilegeResponse |
ThriftHiveMetastore.Client.recv_grant_revoke_privileges() |
GrantRevokeRoleResponse |
ThriftHiveMetastore.Client.recv_grant_revoke_role() |
boolean |
ThriftHiveMetastore.Client.recv_grant_role() |
boolean |
ThriftHiveMetastore.Client.recv_isPartitionMarkedForEvent() |
List<HiveObjectPrivilege> |
ThriftHiveMetastore.Client.recv_list_privileges() |
List<Role> |
ThriftHiveMetastore.Client.recv_list_roles() |
void |
ThriftHiveMetastore.Client.recv_markPartitionForEvent() |
boolean |
ThriftHiveMetastore.Client.recv_partition_name_has_valid_characters() |
Map<String,String> |
ThriftHiveMetastore.Client.recv_partition_name_to_spec() |
List<String> |
ThriftHiveMetastore.Client.recv_partition_name_to_vals() |
void |
ThriftHiveMetastore.Client.recv_rename_partition() |
long |
ThriftHiveMetastore.Client.recv_renew_delegation_token() |
boolean |
ThriftHiveMetastore.Client.recv_revoke_privileges() |
boolean |
ThriftHiveMetastore.Client.recv_revoke_role() |
boolean |
ThriftHiveMetastore.Client.recv_set_aggr_stats_for() |
List<String> |
ThriftHiveMetastore.Client.recv_set_ugi() |
void |
ThriftHiveMetastore.Client.recv_setMetaConf() |
boolean |
ThriftHiveMetastore.Client.recv_update_partition_column_statistics() |
boolean |
ThriftHiveMetastore.Client.recv_update_table_column_statistics() |
void |
ThriftHiveMetastore.Iface.rename_partition(String db_name,
String tbl_name,
List<String> part_vals,
Partition new_part) |
void |
ThriftHiveMetastore.Client.rename_partition(String db_name,
String tbl_name,
List<String> part_vals,
Partition new_part) |
long |
ThriftHiveMetastore.Iface.renew_delegation_token(String token_str_form) |
long |
ThriftHiveMetastore.Client.renew_delegation_token(String token_str_form) |
boolean |
ThriftHiveMetastore.Iface.revoke_privileges(PrivilegeBag privileges) |
boolean |
ThriftHiveMetastore.Client.revoke_privileges(PrivilegeBag privileges) |
boolean |
ThriftHiveMetastore.Iface.revoke_role(String role_name,
String principal_name,
PrincipalType principal_type) |
boolean |
ThriftHiveMetastore.Client.revoke_role(String role_name,
String principal_name,
PrincipalType principal_type) |
boolean |
ThriftHiveMetastore.Iface.set_aggr_stats_for(SetPartitionsStatsRequest request) |
boolean |
ThriftHiveMetastore.Client.set_aggr_stats_for(SetPartitionsStatsRequest request) |
List<String> |
ThriftHiveMetastore.Iface.set_ugi(String user_name,
List<String> group_names) |
List<String> |
ThriftHiveMetastore.Client.set_ugi(String user_name,
List<String> group_names) |
void |
ThriftHiveMetastore.Iface.setMetaConf(String key,
String value) |
void |
ThriftHiveMetastore.Client.setMetaConf(String key,
String value) |
boolean |
ThriftHiveMetastore.Iface.update_partition_column_statistics(ColumnStatistics stats_obj) |
boolean |
ThriftHiveMetastore.Client.update_partition_column_statistics(ColumnStatistics stats_obj) |
boolean |
ThriftHiveMetastore.Iface.update_table_column_statistics(ColumnStatistics stats_obj) |
boolean |
ThriftHiveMetastore.Client.update_table_column_statistics(ColumnStatistics stats_obj) |
Constructor and Description |
---|
InsertEvent(String db,
String table,
List<String> partVals,
List<String> files,
boolean status,
HiveMetaStore.HMSHandler handler) |
Modifier and Type | Method and Description |
---|---|
void |
ExpressionTree.accept(ExpressionTree.TreeVisitor treeVisitor) |
protected void |
ExpressionTree.TreeNode.accept(ExpressionTree.TreeVisitor visitor)
Double dispatch for TreeVisitor.
|
protected void |
ExpressionTree.LeafNode.accept(ExpressionTree.TreeVisitor visitor) |
protected void |
ExpressionTree.TreeVisitor.beginTreeNode(ExpressionTree.TreeNode node) |
protected void |
ExpressionTree.TreeVisitor.endTreeNode(ExpressionTree.TreeNode node) |
void |
ExpressionTree.TreeNode.generateJDOFilter(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuffer)
Generates a JDO filter statement
|
void |
ExpressionTree.LeafNode.generateJDOFilter(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuilder) |
void |
ExpressionTree.generateJDOFilterFragment(org.apache.hadoop.conf.Configuration conf,
Table table,
Map<String,Object> params,
ExpressionTree.FilterBuilder filterBuilder)
Generate the JDOQL filter for the given expression tree
|
String |
ExpressionTree.FilterBuilder.getFilter() |
int |
ExpressionTree.LeafNode.getPartColIndexForFilter(Table table,
ExpressionTree.FilterBuilder filterBuilder)
Get partition column index in the table partition column list that
corresponds to the key that is being filtered on by this tree node.
|
protected void |
ExpressionTree.TreeVisitor.midTreeNode(ExpressionTree.TreeNode node) |
void |
ExpressionTree.FilterBuilder.setError(String errorMessage) |
protected void |
ExpressionTree.TreeVisitor.visit(ExpressionTree.LeafNode node) |
Modifier and Type | Method and Description |
---|---|
void |
PartitionSpecWithSharedSDProxy.setRootLocation(String rootLocation) |
abstract void |
PartitionSpecProxy.setRootLocation(String rootLocation)
Setter for the common root-location for all partitions in the PartitionSet.
|
void |
PartitionListComposingSpecProxy.setRootLocation(String newRootPath) |
void |
CompositePartitionSpecProxy.setRootLocation(String rootLocation) |
Modifier and Type | Method and Description |
---|---|
void |
TxnHandler.abortTxn(AbortTxnRequest rqst) |
void |
TxnHandler.addDynamicPartitions(AddDynamicPartitions rqst) |
LockResponse |
TxnHandler.checkLock(CheckLockRequest rqst) |
protected void |
TxnHandler.checkRetryable(Connection conn,
SQLException e,
String caller)
Determine if an exception was such that it makse sense to retry.
|
void |
CompactionTxnHandler.cleanEmptyAbortedTxns()
Clean up aborted transactions from txns that have no components in txn_components.
|
void |
TxnHandler.commitTxn(CommitTxnRequest rqst) |
void |
TxnHandler.compact(CompactionRequest rqst) |
protected TxnHandler.DatabaseProduct |
TxnHandler.determineDatabaseProduct(Connection conn)
Determine the database product type
|
List<String> |
CompactionTxnHandler.findColumnsWithStats(CompactionInfo ci)
Queries metastore DB directly to find columns in the table which have statistics information.
|
CompactionInfo |
CompactionTxnHandler.findNextToCompact(String workerId)
This will grab the next compaction request off of
the queue, and assign it to the worker.
|
Set<CompactionInfo> |
CompactionTxnHandler.findPotentialCompactions(int maxAborted)
This will look through the completed_txn_components table and look for partitions or tables
that may be ready for compaction.
|
List<CompactionInfo> |
CompactionTxnHandler.findReadyToClean()
Find entries in the queue that are ready to
be cleaned.
|
protected long |
TxnHandler.getDbTime(Connection conn)
Determine the current time, using the RDBMS as a source of truth
|
GetOpenTxnsResponse |
TxnHandler.getOpenTxns() |
GetOpenTxnsInfoResponse |
TxnHandler.getOpenTxnsInfo() |
void |
TxnHandler.heartbeat(HeartbeatRequest ids) |
HeartbeatTxnRangeResponse |
TxnHandler.heartbeatTxnRange(HeartbeatTxnRangeRequest rqst) |
LockResponse |
TxnHandler.lock(LockRequest rqst) |
LockResponse |
TxnHandler.lockNoWait(LockRequest rqst) |
void |
CompactionTxnHandler.markCleaned(CompactionInfo info)
This will remove an entry from the queue after
it has been compacted.
|
void |
CompactionTxnHandler.markCompacted(CompactionInfo info)
This will mark an entry in the queue as compacted
and put it in the ready to clean state.
|
OpenTxnsResponse |
TxnHandler.openTxns(OpenTxnRequest rqst) |
void |
CompactionTxnHandler.revokeFromLocalWorkers(String hostname)
This will take all entries assigned to workers
on a host return them to INITIATED state.
|
void |
CompactionTxnHandler.revokeTimedoutWorkers(long timeout)
This call will return all compaction queue
entries assigned to a worker but over the timeout back to the initiated state.
|
void |
CompactionTxnHandler.setRunAs(long cq_id,
String user)
Sets the user to run as.
|
ShowCompactResponse |
TxnHandler.showCompact(ShowCompactRequest rqst) |
ShowLocksResponse |
TxnHandler.showLocks(ShowLocksRequest rqst) |
void |
TxnHandler.unlock(UnlockRequest rqst) |
Modifier and Type | Method and Description |
---|---|
void |
SessionHiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
boolean cascade) |
void |
SessionHiveMetaStoreClient.alter_table(String dbname,
String tbl_name,
Table new_tbl,
EnvironmentContext envContext) |
protected void |
SessionHiveMetaStoreClient.create_table_with_environment_context(Table tbl,
EnvironmentContext envContext) |
boolean |
SessionHiveMetaStoreClient.deleteTableColumnStatistics(String dbName,
String tableName,
String colName)
Delete table level column statistics given dbName, tableName and colName
|
protected void |
SessionHiveMetaStoreClient.drop_table_with_environment_context(String dbname,
String name,
boolean deleteData,
EnvironmentContext envContext) |
PrincipalPrivilegeSet |
SessionHiveMetaStoreClient.get_privilege_set(HiveObjectRef hiveObject,
String userName,
List<String> groupNames) |
List<String> |
SessionHiveMetaStoreClient.getAllTables(String dbName) |
IMetaStoreClient |
Hive.getMSC() |
List<Partition> |
Hive.getPartitionsByFilter(Table tbl,
String filter)
Get a list of Partitions by filter.
|
List<FieldSchema> |
SessionHiveMetaStoreClient.getSchema(String dbName,
String tableName) |
Table |
SessionHiveMetaStoreClient.getTable(String dbname,
String name) |
List<ColumnStatisticsObj> |
SessionHiveMetaStoreClient.getTableColumnStatistics(String dbName,
String tableName,
List<String> colNames)
Get table column statistics given dbName, tableName and multiple colName-s
|
List<Table> |
SessionHiveMetaStoreClient.getTableObjectsByName(String dbName,
List<String> tableNames) |
List<String> |
SessionHiveMetaStoreClient.getTables(String dbName,
String tablePattern) |
boolean |
SessionHiveMetaStoreClient.tableExists(String databaseName,
String tableName) |
boolean |
SessionHiveMetaStoreClient.updateTableColumnStatistics(ColumnStatistics statsObj)
Write table level column statistics to persistent store
|
Modifier and Type | Method and Description |
---|---|
String |
PartitionExpressionForMetastore.convertExprToFilter(byte[] exprBytes) |
boolean |
PartitionExpressionForMetastore.filterPartitionsByExpr(List<String> partColumnNames,
List<PrimitiveTypeInfo> partColumnTypeInfos,
byte[] exprBytes,
String defaultPartitionName,
List<String> partitionNames) |
static boolean |
PartitionPruner.prunePartitionNames(List<String> partColumnNames,
List<PrimitiveTypeInfo> partColumnTypeInfos,
ExprNodeGenericFuncDesc prunerExpr,
String defaultPartitionName,
List<String> partNames)
Prunes partition names to see if they match the prune expression.
|
Modifier and Type | Method and Description |
---|---|
void |
MetaDataExportListener.onEvent(PreEventContext context)
Listen for an event; if it is a DROP_TABLE event, call export_meta_data
|
Modifier and Type | Method and Description |
---|---|
void |
AuthorizationPreEventListener.onEvent(PreEventContext context) |
Constructor and Description |
---|
AuthorizationPreEventListener.PartitionWrapper(Partition mapiPart,
PreEventContext context) |
Modifier and Type | Method and Description |
---|---|
List<String> |
AuthorizationMetaStoreFilterHook.filterDatabases(List<String> dbList) |
List<String> |
AuthorizationMetaStoreFilterHook.filterTableNames(String dbName,
List<String> tableList) |
Modifier and Type | Method and Description |
---|---|
void |
Worker.init(AtomicBoolean stop,
AtomicBoolean looped) |
void |
Initiator.init(AtomicBoolean stop,
AtomicBoolean looped) |
Modifier and Type | Method and Description |
---|---|
static org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier> |
HCatUtil.extractThriftToken(String tokenStrForm,
String tokenSignature) |
static HiveMetaStoreClient |
HCatUtil.getHiveClient(HiveConf hiveConf)
Deprecated.
|
static IMetaStoreClient |
HCatUtil.getHiveMetastoreClient(HiveConf hiveConf)
Get or create a hive client depending on whether it exits in cache or not
|
static Table |
HCatUtil.getTable(IMetaStoreClient client,
String dbName,
String tableName) |
Modifier and Type | Method and Description |
---|---|
void |
NotificationListener.onAddPartition(AddPartitionEvent partitionEvent) |
void |
DbNotificationListener.onAddPartition(AddPartitionEvent partitionEvent) |
void |
NotificationListener.onAlterPartition(AlterPartitionEvent ape) |
void |
DbNotificationListener.onAlterPartition(AlterPartitionEvent partitionEvent) |
void |
NotificationListener.onAlterTable(AlterTableEvent tableEvent)
Send altered table notifications.
|
void |
DbNotificationListener.onAlterTable(AlterTableEvent tableEvent) |
void |
DbNotificationListener.onConfigChange(ConfigChangeEvent tableEvent) |
void |
NotificationListener.onCreateDatabase(CreateDatabaseEvent dbEvent) |
void |
DbNotificationListener.onCreateDatabase(CreateDatabaseEvent dbEvent) |
void |
NotificationListener.onCreateTable(CreateTableEvent tableEvent) |
void |
DbNotificationListener.onCreateTable(CreateTableEvent tableEvent) |
void |
NotificationListener.onDropDatabase(DropDatabaseEvent dbEvent) |
void |
DbNotificationListener.onDropDatabase(DropDatabaseEvent dbEvent) |
void |
NotificationListener.onDropPartition(DropPartitionEvent partitionEvent)
Send dropped partition notifications.
|
void |
DbNotificationListener.onDropPartition(DropPartitionEvent partitionEvent) |
void |
NotificationListener.onDropTable(DropTableEvent tableEvent)
Send dropped table notifications.
|
void |
DbNotificationListener.onDropTable(DropTableEvent tableEvent) |
void |
DbNotificationListener.onInsert(InsertEvent insertEvent) |
void |
NotificationListener.onLoadPartitionDone(LoadPartitionDoneEvent lpde) |
void |
DbNotificationListener.onLoadPartitionDone(LoadPartitionDoneEvent partSetDoneEvent) |
Copyright © 2017 The Apache Software Foundation. All rights reserved.