Package | Description |
---|---|
org.apache.hadoop.hive.ql.exec |
Hive QL execution tasks, operators, functions and other handlers.
|
org.apache.hadoop.hive.ql.exec.repl.bootstrap.load.util | |
org.apache.hadoop.hive.ql.io.orc |
The Optimized Row Columnar (ORC) File Format.
|
org.apache.hadoop.hive.ql.lockmgr |
Hive Lock Manager interfaces and some custom implmentations
|
org.apache.hadoop.hive.ql.metadata | |
org.apache.hadoop.hive.ql.metadata.formatting | |
org.apache.hadoop.hive.ql.parse | |
org.apache.hadoop.hive.ql.parse.authorization | |
org.apache.hadoop.hive.ql.parse.repl.dump | |
org.apache.hadoop.hive.ql.parse.repl.dump.events | |
org.apache.hadoop.hive.ql.parse.repl.load.message | |
org.apache.hadoop.hive.ql.plan | |
org.apache.hadoop.hive.ql.security.authorization | |
org.apache.hadoop.hive.ql.security.authorization.plugin | |
org.apache.hadoop.hive.ql.stats | |
org.apache.hive.hcatalog.cli.SemanticAnalysis | |
org.apache.hive.service.cli.session |
Modifier and Type | Method and Description |
---|---|
protected Hive |
Task.getHive() |
Modifier and Type | Method and Description |
---|---|
static String |
ArchiveUtils.conflictingArchiveNameOrNull(Hive db,
Table tbl,
LinkedHashMap<String,String> partSpec)
Determines if one can insert into partition(s), or there's a conflict with
archive.
|
int |
DDLTask.showColumns(Hive db,
ShowColumnsDesc showCols)
Write a list of the columns in the table to a file.
|
Modifier and Type | Field and Description |
---|---|
Hive |
Context.hiveDb |
Constructor and Description |
---|
Context(String dumpDirectory,
HiveConf hiveConf,
Hive hiveDb,
LineageState lineageState,
Context nestedContext) |
Constructor and Description |
---|
HBaseCache(Hive hive) |
Modifier and Type | Method and Description |
---|---|
static HiveLockObject |
HiveLockObject.createFrom(Hive hiveDB,
String tableName,
Map<String,String> partSpec)
Creates a locking object for a table (when partition spec is not provided)
or a table partition
|
int |
DbTxnManager.lockDatabase(Hive hiveDB,
LockDatabaseDesc lockDb) |
int |
HiveTxnManager.lockDatabase(Hive hiveDB,
LockDatabaseDesc lockDb)
This function is called to lock the database when explicit lock command is
issued on a database.
|
int |
DbTxnManager.lockTable(Hive db,
LockTableDesc lockTbl) |
int |
HiveTxnManager.lockTable(Hive hiveDB,
LockTableDesc lockTbl)
This function is called to lock the table when explicit lock command is
issued on a table.
|
int |
DbTxnManager.unlockDatabase(Hive hiveDB,
UnlockDatabaseDesc unlockDb) |
int |
HiveTxnManager.unlockDatabase(Hive hiveDB,
UnlockDatabaseDesc unlockDb)
This function is called to unlock the database when explicit unlock command
is issued on a database.
|
int |
DbTxnManager.unlockTable(Hive hiveDB,
UnlockTableDesc unlockTbl) |
int |
HiveTxnManager.unlockTable(Hive hiveDB,
UnlockTableDesc unlockTbl)
This function is called to unlock the table when explicit unlock command is
issued on a table.
|
Modifier and Type | Method and Description |
---|---|
static Hive |
Hive.get() |
static Hive |
Hive.get(boolean doRegisterAllFns) |
static Hive |
Hive.get(org.apache.hadoop.conf.Configuration c,
Class<?> clazz) |
static Hive |
Hive.get(HiveConf c)
Gets hive object for the current thread.
|
static Hive |
Hive.get(HiveConf c,
boolean needsRefresh)
get a connection to metastore.
|
static Hive |
Hive.getWithFastCheck(HiveConf c)
Same as
get(HiveConf) , except that it checks only the object identity of existing
MS client, assuming the relevant settings would be unchanged within the same conf object. |
static Hive |
Hive.getWithFastCheck(HiveConf c,
boolean doRegisterAllFns)
Same as
get(HiveConf) , except that it checks only the object identity of existing
MS client, assuming the relevant settings would be unchanged within the same conf object. |
static Hive |
Hive.getWithoutRegisterFns(HiveConf c)
Same as
get(HiveConf) , except that it does not register all functions. |
Modifier and Type | Method and Description |
---|---|
void |
HiveMaterializedViewsRegistry.init(Hive db) |
static void |
Hive.set(Hive hive) |
Constructor and Description |
---|
HiveMetaStoreChecker(Hive hive) |
PartitionIterable(Hive db,
Table table,
Map<String,String> partialPartitionSpec,
int batch_size)
Primary constructor that fetches all partitions in a given table, given
a Hive object and a table object, and a partial partition spec.
|
Modifier and Type | Method and Description |
---|---|
void |
MetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par)
Show the table status.
|
void |
JsonMetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par) |
Modifier and Type | Field and Description |
---|---|
protected Hive |
TaskCompiler.db |
protected Hive |
BaseSemanticAnalyzer.db |
Modifier and Type | Method and Description |
---|---|
protected static Hive |
BaseSemanticAnalyzer.createHiveDB(HiveConf conf) |
Hive |
BaseSemanticAnalyzer.getDb() |
Hive |
HiveSemanticAnalyzerHookContextImpl.getHive() |
Hive |
HiveSemanticAnalyzerHookContext.getHive() |
Hive |
EximUtil.SemanticAnalyzerWrapperContext.getHive() |
Modifier and Type | Method and Description |
---|---|
void |
TaskCompiler.init(QueryState queryState,
SessionState.LogHelper console,
Hive db) |
void |
TezCompiler.init(QueryState queryState,
SessionState.LogHelper console,
Hive db) |
protected static ASTNode |
SemanticAnalyzer.rewriteASTWithMaskAndFilter(TableMask tableMask,
ASTNode ast,
org.antlr.runtime.TokenRewriteStream tokenRewriteStream,
Context ctx,
Hive db,
Map<String,Table> tabNameToTabObject,
Set<Integer> ignoredTokens) |
static Table |
ImportSemanticAnalyzer.tableIfExists(ImportTableDesc tblDesc,
Hive db)
Utility method that returns a table if one corresponding to the destination
tblDesc is found.
|
Constructor and Description |
---|
BaseSemanticAnalyzer(QueryState queryState,
Hive db) |
DDLSemanticAnalyzer(QueryState queryState,
Hive db) |
SemanticAnalyzerWrapperContext(HiveConf conf,
Hive db,
HashSet<ReadEntity> inputs,
HashSet<WriteEntity> outputs,
List<Task<? extends Serializable>> tasks,
org.slf4j.Logger LOG,
Context ctx) |
TableSpec(Hive db,
HiveConf conf,
ASTNode ast) |
TableSpec(Hive db,
HiveConf conf,
ASTNode ast,
boolean allowDynamicPartitionsSpec,
boolean allowPartialPartitionsSpec) |
TableSpec(Hive db,
String tableName,
Map<String,String> partSpec) |
TableSpec(Hive db,
String tableName,
Map<String,String> partSpec,
boolean allowPartialPartitionsSpec) |
Constructor and Description |
---|
HiveAuthorizationTaskFactoryImpl(HiveConf conf,
Hive db) |
Modifier and Type | Method and Description |
---|---|
static Collection<String> |
Utils.getAllTables(Hive db,
String dbName) |
static boolean |
Utils.isBootstrapDumpInProgress(Hive hiveDb,
String dbName) |
static Iterable<? extends String> |
Utils.matchesDb(Hive db,
String dbPattern) |
static Iterable<? extends String> |
Utils.matchesTbl(Hive db,
String dbName,
String tblPattern) |
static void |
Utils.resetDbBootstrapDumpState(Hive hiveDb,
String dbName,
String uniqueKey) |
static String |
Utils.setDbBootstrapDumpState(Hive hiveDb,
String dbName) |
static boolean |
Utils.shouldReplicate(NotificationEvent tableForEvent,
ReplicationSpec replicationSpec,
Hive db,
HiveConf hiveConf) |
Constructor and Description |
---|
HiveWrapper(Hive db,
String dbName) |
HiveWrapper(Hive db,
String dbName,
long lastReplId) |
TableExport(TableExport.Paths paths,
BaseSemanticAnalyzer.TableSpec tableSpec,
ReplicationSpec replicationSpec,
Hive db,
String distCpDoAsUser,
HiveConf conf,
ExportWork.MmContext mmCtx) |
Constructor and Description |
---|
Context(org.apache.hadoop.fs.Path eventRoot,
org.apache.hadoop.fs.Path cmRoot,
Hive db,
HiveConf hiveConf,
ReplicationSpec replicationSpec) |
Constructor and Description |
---|
Context(String dbName,
String tableName,
String location,
Task<? extends Serializable> precursor,
DumpMetaData dmd,
HiveConf hiveConf,
Hive db,
Context nestedContext,
org.slf4j.Logger log) |
Modifier and Type | Method and Description |
---|---|
void |
ExportWork.acidPostProcess(Hive db)
For exporting Acid table, change the "pointer" to the temp table.
|
Constructor and Description |
---|
HiveProxy(Hive hive) |
Constructor and Description |
---|
HiveV1Authorizer(HiveConf conf,
Hive hive)
Deprecated.
|
Modifier and Type | Method and Description |
---|---|
int |
ColStatsProcessor.persistColumnStats(Hive db,
Table tbl) |
int |
BasicStatsTask.process(Hive db,
Table tbl) |
int |
BasicStatsNoJobTask.process(Hive db,
Table tbl) |
int |
IStatsProcessor.process(Hive db,
Table tbl) |
int |
ColStatsProcessor.process(Hive db,
Table tbl) |
Modifier and Type | Method and Description |
---|---|
protected void |
HCatSemanticAnalyzerBase.authorizeDDLWork(HiveSemanticAnalyzerHookContext context,
Hive hive,
DDLWork work)
Authorized the given DDLWork.
|
protected void |
HCatSemanticAnalyzer.authorizeDDLWork(HiveSemanticAnalyzerHookContext cntxt,
Hive hive,
DDLWork work) |
protected void |
HCatSemanticAnalyzerBase.authorizeTable(Hive hive,
String tableName,
Privilege priv) |
Modifier and Type | Method and Description |
---|---|
Hive |
HiveSessionImpl.getSessionHive() |
Hive |
HiveSession.getSessionHive() |
Copyright © 2022 The Apache Software Foundation. All rights reserved.