Package | Description |
---|---|
org.apache.hadoop.hive.ql.exec |
Hive QL execution tasks, operators, functions and other handlers.
|
org.apache.hadoop.hive.ql.hooks | |
org.apache.hadoop.hive.ql.index | |
org.apache.hadoop.hive.ql.lockmgr |
Hive Lock Manager interfaces and some custom implmentations
|
org.apache.hadoop.hive.ql.metadata | |
org.apache.hadoop.hive.ql.metadata.formatting | |
org.apache.hadoop.hive.ql.optimizer | |
org.apache.hadoop.hive.ql.optimizer.calcite | |
org.apache.hadoop.hive.ql.optimizer.metainfo.annotation | |
org.apache.hadoop.hive.ql.optimizer.ppr | |
org.apache.hadoop.hive.ql.parse | |
org.apache.hadoop.hive.ql.plan | |
org.apache.hadoop.hive.ql.security.authorization | |
org.apache.hadoop.hive.ql.session | |
org.apache.hadoop.hive.ql.stats | |
org.apache.hive.hcatalog.cli.SemanticAnalysis | |
org.apache.hive.hcatalog.common |
Modifier and Type | Method and Description |
---|---|
static String |
ArchiveUtils.conflictingArchiveNameOrNull(Hive db,
Table tbl,
LinkedHashMap<String,String> partSpec)
Determines if one can insert into partition(s), or there's a conflict with
archive.
|
static ArchiveUtils.PartSpecInfo |
ArchiveUtils.PartSpecInfo.create(Table tbl,
Map<String,String> partSpec)
Extract partial prefix specification from table and key-value map
|
org.apache.hadoop.fs.Path |
ArchiveUtils.PartSpecInfo.createPath(Table tbl)
Creates path where partitions matching prefix should lie in filesystem
|
static TableDesc |
Utilities.getTableDesc(Table tbl) |
Modifier and Type | Method and Description |
---|---|
Table |
Entity.getT() |
Table |
Entity.getTable()
Get the table associated with the entity.
|
Modifier and Type | Method and Description |
---|---|
void |
Entity.setT(Table t) |
Constructor and Description |
---|
Entity(Table t,
boolean complete)
Constructor for a table.
|
ReadEntity(Table t)
Constructor.
|
ReadEntity(Table t,
ReadEntity parent) |
ReadEntity(Table t,
ReadEntity parent,
boolean isDirect) |
WriteEntity(Table t,
WriteEntity.WriteType type)
Constructor for a table.
|
WriteEntity(Table t,
WriteEntity.WriteType type,
boolean complete) |
Modifier and Type | Method and Description |
---|---|
List<Task<?>> |
TableBasedIndexHandler.generateIndexBuildTaskList(Table baseTbl,
Index index,
List<Partition> indexTblPartitions,
List<Partition> baseTblPartitions,
Table indexTbl,
Set<ReadEntity> inputs,
Set<WriteEntity> outputs) |
List<Task<?>> |
HiveIndexHandler.generateIndexBuildTaskList(Table baseTbl,
Index index,
List<Partition> indexTblPartitions,
List<Partition> baseTblPartitions,
Table indexTbl,
Set<ReadEntity> inputs,
Set<WriteEntity> outputs)
Requests that the handler generate a plan for building the index; the plan
should read the base table and write out the index representation.
|
Constructor and Description |
---|
HiveLockObject(Table tbl,
HiveLockObject.HiveLockObjectData lockData) |
Modifier and Type | Method and Description |
---|---|
Table |
Table.copy() |
Table |
Partition.getTable() |
Table |
Hive.getTable(String tableName)
Returns metadata for the table named tableName
|
Table |
Hive.getTable(String tableName,
boolean throwException)
Returns metadata for the table named tableName
|
Table |
Hive.getTable(String dbName,
String tableName)
Returns metadata of the table
|
Table |
Hive.getTable(String dbName,
String tableName,
boolean throwException)
Returns metadata of the table
|
Table |
Hive.newTable(String tableName) |
Modifier and Type | Method and Description |
---|---|
void |
Hive.alterTable(String tblName,
Table newTbl)
Updates the existing table metadata with the new metadata.
|
void |
Hive.alterTable(String tblName,
Table newTbl,
boolean cascade) |
static StorageDescriptor |
Partition.cloneSd(Table tbl)
We already have methods that clone stuff using XML or Kryo.
|
static Partition |
Partition.createMetaPartitionObject(Table tbl,
Map<String,String> partSpec,
org.apache.hadoop.fs.Path location) |
Partition |
Hive.createPartition(Table tbl,
Map<String,String> partSpec)
Creates a partition.
|
void |
Hive.createTable(Table tbl)
Creates the table with the give objects
|
void |
Hive.createTable(Table tbl,
boolean ifNotExists)
Creates the table with the give objects
|
Set<Partition> |
Hive.getAllPartitionsOf(Table tbl)
Get all the partitions; unlike
Hive.getPartitions(Table) , does not include auth. |
Partition |
Hive.getPartition(Table tbl,
Map<String,String> partSpec,
boolean forceCreate) |
Partition |
Hive.getPartition(Table tbl,
Map<String,String> partSpec,
boolean forceCreate,
String partPath,
boolean inheritTableSpecs)
Returns partition metadata
|
Partition |
Hive.getPartition(Table tbl,
Map<String,String> partSpec,
boolean forceCreate,
String partPath,
boolean inheritTableSpecs,
List<org.apache.hadoop.fs.Path> newFiles)
Returns partition metadata
|
List<Partition> |
Hive.getPartitions(Table tbl)
get all the partitions that the table has
|
List<Partition> |
Hive.getPartitions(Table tbl,
Map<String,String> partialPartSpec)
get all the partitions of the table that matches the given partial
specification.
|
List<Partition> |
Hive.getPartitions(Table tbl,
Map<String,String> partialPartSpec,
short limit)
get all the partitions of the table that matches the given partial
specification.
|
boolean |
Hive.getPartitionsByExpr(Table tbl,
ExprNodeGenericFuncDesc expr,
HiveConf conf,
List<Partition> result)
Get a list of Partitions by expr.
|
List<Partition> |
Hive.getPartitionsByFilter(Table tbl,
String filter)
Get a list of Partitions by filter.
|
List<Partition> |
Hive.getPartitionsByNames(Table tbl,
List<String> partNames)
Get all partitions of the table that matches the list of given partition names.
|
List<Partition> |
Hive.getPartitionsByNames(Table tbl,
Map<String,String> partialPartSpec)
get all the partitions of the table that matches the given partial
specification.
|
protected void |
Partition.initialize(Table table,
Partition tPartition)
Initializes this object with the given variables
|
Partition |
Hive.loadPartition(org.apache.hadoop.fs.Path loadPath,
Table tbl,
Map<String,String> partSpec,
boolean replace,
boolean holdDDLTime,
boolean inheritTableSpecs,
boolean isSkewedStoreAsSubdir,
boolean isSrcLocal,
boolean isAcid)
Load a directory into a Hive Table Partition - Alters existing content of
the partition with the contents of loadPath.
|
void |
Hive.renamePartition(Table tbl,
Map<String,String> oldPartSpec,
Partition newPart)
Rename a old partition to new partition
|
void |
Partition.setTable(Table table)
Should be only used by serialization.
|
Constructor and Description |
---|
DummyPartition(Table tbl,
String name) |
DummyPartition(Table tbl,
String name,
Map<String,String> partSpec) |
Partition(Table tbl)
create an empty partition.
|
Partition(Table tbl,
Map<String,String> partSpec,
org.apache.hadoop.fs.Path location)
Create partition object with the given info.
|
Partition(Table tbl,
Partition tp) |
PartitionIterable(Hive db,
Table table,
Map<String,String> partialPartitionSpec,
int batch_size)
Primary constructor that fetches all partitions in a given table, given
a Hive object and a table object, and a partial partition spec.
|
Modifier and Type | Method and Description |
---|---|
void |
MetaDataFormatter.describeTable(DataOutputStream out,
String colPath,
String tableName,
Table tbl,
Partition part,
List<FieldSchema> cols,
boolean isFormatted,
boolean isExt,
boolean isPretty,
boolean isOutputPadded,
List<ColumnStatisticsObj> colStats)
Describe table.
|
void |
JsonMetaDataFormatter.describeTable(DataOutputStream out,
String colPath,
String tableName,
Table tbl,
Partition part,
List<FieldSchema> cols,
boolean isFormatted,
boolean isExt,
boolean isPretty,
boolean isOutputPadded,
List<ColumnStatisticsObj> colStats)
Describe table.
|
static String |
MetaDataFormatUtils.getTableInformation(Table table) |
Modifier and Type | Method and Description |
---|---|
void |
MetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par)
Show the table status.
|
void |
JsonMetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par) |
Modifier and Type | Method and Description |
---|---|
static List<Index> |
IndexUtils.getAllIndexes(Table table,
short max) |
static List<Index> |
IndexUtils.getIndexes(Table baseTableMetaData,
List<String> matchIndexTypes)
Get a list of indexes on a table that match given types.
|
protected long |
SizeBasedBigTableSelectorForAutoSMJ.getSize(HiveConf conf,
Table table) |
Modifier and Type | Method and Description |
---|---|
Table |
RelOptHiveTable.getHiveTableMD() |
Constructor and Description |
---|
RelOptHiveTable(org.apache.calcite.plan.RelOptSchema calciteSchema,
String qualifiedTblName,
org.apache.calcite.rel.type.RelDataType rowType,
Table hiveTblMetadata,
List<ColumnInfo> hiveNonPartitionCols,
List<ColumnInfo> hivePartitionCols,
List<VirtualColumn> hiveVirtualCols,
HiveConf hconf,
Map<String,PrunedPartitionList> partitionCache,
AtomicInteger noColsMissingStats) |
Modifier and Type | Method and Description |
---|---|
boolean |
OpTraitsRulesProcFactory.TableScanRule.checkBucketedTable(Table tbl,
ParseContext pGraphContext,
PrunedPartitionList prunedParts) |
Modifier and Type | Method and Description |
---|---|
static boolean |
PartitionPruner.onlyContainsPartnCols(Table tab,
ExprNodeDesc expr)
Find out whether the condition only contains partitioned columns.
|
static PrunedPartitionList |
PartitionPruner.prune(Table tab,
ExprNodeDesc prunerExpr,
HiveConf conf,
String alias,
Map<String,PrunedPartitionList> prunedPartitionsMap)
Get the partition list for the table that satisfies the partition pruner
condition.
|
Modifier and Type | Field and Description |
---|---|
Table |
BaseSemanticAnalyzer.TableSpec.tableHandle |
Modifier and Type | Method and Description |
---|---|
Table |
QBMetaData.getDestTableForAlias(String alias) |
Table |
PrunedPartitionList.getSourceTable() |
Table |
QBMetaData.getSrcForAlias(String alias) |
protected Table |
BaseSemanticAnalyzer.getTable(String tblName) |
protected Table |
BaseSemanticAnalyzer.getTable(String[] qualified) |
protected Table |
BaseSemanticAnalyzer.getTable(String[] qualified,
boolean throwException) |
protected Table |
BaseSemanticAnalyzer.getTable(String tblName,
boolean throwException) |
protected Table |
BaseSemanticAnalyzer.getTable(String database,
String tblName,
boolean throwException) |
Table |
SemanticAnalyzer.getTable(TableScanOperator ts) |
Table |
QBMetaData.getTableForAlias(String alias) |
Modifier and Type | Method and Description |
---|---|
HashMap<String,Table> |
QBMetaData.getAliasToTable() |
Map<String,Table> |
QBMetaData.getNameToDestTable() |
Modifier and Type | Method and Description |
---|---|
boolean |
ReplicationSpec.allowEventReplacementInto(Table table)
Determines if a current replication event specification is allowed to
replicate-replace-into a given table
|
boolean |
ReplicationSpec.allowReplacementInto(Table table)
Determines if a current replication object(current state of dump) is allowed to
replicate-replace-into a given table
|
static void |
EximUtil.createExportDump(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path metadataPath,
Table tableHandle,
Iterable<Partition> partitions,
ReplicationSpec replicationSpec) |
protected Partition |
BaseSemanticAnalyzer.getPartition(Table table,
Map<String,String> partSpec,
boolean throwException) |
protected List<Partition> |
BaseSemanticAnalyzer.getPartitions(Table table,
Map<String,String> partSpec,
boolean throwException) |
static HashMap<String,String> |
DDLSemanticAnalyzer.getValidatedPartSpec(Table table,
ASTNode astNode,
HiveConf conf,
boolean shouldBeFull) |
static boolean |
SemanticAnalyzer.isAcidTable(Table tab) |
static boolean |
DDLSemanticAnalyzer.isFullSpec(Table table,
Map<String,String> partSpec) |
boolean |
BaseSemanticAnalyzer.isValidPrefixSpec(Table tTable,
Map<String,String> spec)
Checks if given specification is proper specification for prefix of
partition cols, for table partitioned by ds, hr, min valid ones are
(ds='2008-04-08'), (ds='2008-04-08', hr='12'), (ds='2008-04-08', hr='12', min='30')
invalid one is for example (ds='2008-04-08', min='30')
|
void |
QBMetaData.setDestForAlias(String alias,
Table tab) |
void |
QBMetaData.setSrcForAlias(String alias,
Table tab) |
static void |
BaseSemanticAnalyzer.validatePartColumnType(Table tbl,
Map<String,String> partSpec,
ASTNode astNode,
HiveConf conf) |
static void |
BaseSemanticAnalyzer.validatePartSpec(Table tbl,
Map<String,String> partSpec,
ASTNode astNode,
HiveConf conf,
boolean shouldBeFull) |
Constructor and Description |
---|
PrunedPartitionList(Table source,
Set<Partition> partitions,
List<String> referred,
boolean hasUnknowns) |
Modifier and Type | Method and Description |
---|---|
Table |
AlterTableExchangePartition.getDestinationTable() |
Table |
AlterTableExchangePartition.getSourceTable() |
Table |
FileSinkDesc.getTable() |
Table |
AlterTableDesc.getTable() |
Table |
TableScanDesc.getTableMetadata() |
Modifier and Type | Method and Description |
---|---|
void |
AlterTableExchangePartition.setDestinationTable(Table destinationTable) |
void |
AlterTableExchangePartition.setSourceTable(Table sourceTable) |
void |
FileSinkDesc.setTable(Table table) |
void |
AlterTableDesc.setTable(Table table) |
Constructor and Description |
---|
AlterTableExchangePartition(Table sourceTable,
Table destinationTable,
Map<String,String> partitionSpecs) |
DynamicPartitionCtx(Table tbl,
Map<String,String> partSpec,
String defaultPartName,
int maxParts) |
TableScanDesc(String alias,
List<VirtualColumn> vcs,
Table tblMetadata) |
TableScanDesc(String alias,
Table tblMetadata) |
TableScanDesc(Table tblMetadata) |
Modifier and Type | Class and Description |
---|---|
static class |
AuthorizationPreEventListener.TableWrapper |
Modifier and Type | Method and Description |
---|---|
abstract void |
HiveMultiPartitionAuthorizationProviderBase.authorize(Table table,
Iterable<Partition> partitions,
Privilege[] requiredReadPrivileges,
Privilege[] requiredWritePrivileges)
Authorization method for partition sets.
|
void |
StorageBasedAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv) |
void |
MetaStoreAuthzAPIAuthorizerEmbedOnly.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv) |
void |
HiveAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a list of columns.
|
void |
BitSetCheckedAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv) |
void |
StorageBasedAuthorizationProvider.authorize(Table table,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv) |
void |
MetaStoreAuthzAPIAuthorizerEmbedOnly.authorize(Table table,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv) |
void |
HiveAuthorizationProvider.authorize(Table table,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a hive table object.
|
void |
BitSetCheckedAuthorizationProvider.authorize(Table table,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv) |
Constructor and Description |
---|
AuthorizationPreEventListener.PartitionWrapper(Table table,
Partition mapiPart) |
Modifier and Type | Method and Description |
---|---|
Map<String,Map<String,Table>> |
SessionState.getTempTables() |
Modifier and Type | Method and Description |
---|---|
static Statistics |
StatsUtils.collectStatistics(HiveConf conf,
PrunedPartitionList partList,
Table table,
List<ColumnInfo> schema,
List<String> neededColumns,
List<String> referencedColumns,
boolean fetchColStats,
boolean fetchPartStats) |
static Statistics |
StatsUtils.collectStatistics(HiveConf conf,
PrunedPartitionList partList,
Table table,
TableScanOperator tableScanOperator)
Collect table, partition and column level statistics
|
static List<Long> |
StatsUtils.getBasicStatForPartitions(Table table,
List<Partition> parts,
String statType)
Get basic stats of partitions
|
static long |
StatsUtils.getBasicStatForTable(Table table,
String statType)
Get basic stats of table
|
static long |
StatsUtils.getFileSizeForTable(HiveConf conf,
Table table)
Find the bytes on disk occupied by a table
|
static long |
StatsUtils.getNumRows(Table table)
Get number of rows of a give table
|
static long |
StatsUtils.getRawDataSize(Table table)
Get raw data size of a give table
|
static List<ColStatistics> |
StatsUtils.getTableColumnStats(Table table,
List<ColumnInfo> schema,
List<String> neededColumns)
Get table level column statistics from metastore for needed columns
|
static long |
StatsUtils.getTotalSize(Table table)
Get total size of a give table
|
Modifier and Type | Method and Description |
---|---|
protected void |
HCatSemanticAnalyzerBase.authorize(Table table,
Privilege priv) |
Modifier and Type | Method and Description |
---|---|
static Table |
HCatUtil.getTable(IMetaStoreClient client,
String dbName,
String tableName) |
Modifier and Type | Method and Description |
---|---|
static HCatSchema |
HCatUtil.extractSchema(Table table) |
static HCatSchema |
HCatUtil.getPartitionColumns(Table table)
return the partition columns from a table instance
|
static HCatSchema |
HCatUtil.getTableSchemaWithPtnCols(Table table) |
static List<FieldSchema> |
HCatUtil.validatePartitionSchema(Table table,
HCatSchema partitionSchema)
Validate partition schema, checks if the column types match between the
partition and the existing table schema.
|
Copyright © 2017 The Apache Software Foundation. All rights reserved.