public class SparkHashTableSinkOperator extends TerminalOperator<SparkHashTableSinkDesc> implements Serializable
Operator.OperatorFunc, Operator.State
Modifier and Type | Field and Description |
---|---|
protected static org.apache.commons.logging.Log |
LOG |
alias, childOperators, childOperatorsArray, childOperatorsTag, colExprMap, conf, CONTEXT_NAME_KEY, done, groupKeyObject, HIVECOUNTERCREATEDFILES, HIVECOUNTERFATAL, id, inputObjInspectors, isLogDebugEnabled, isLogInfoEnabled, isLogTraceEnabled, operatorId, out, outputObjInspector, parentOperators, PLOG, reporter, state, statsMap
Constructor and Description |
---|
SparkHashTableSinkOperator() |
Modifier and Type | Method and Description |
---|---|
void |
closeOp(boolean abort)
Operator specific close routine.
|
protected void |
flushToFile(MapJoinPersistableTableContainer tableContainer,
byte tag) |
String |
getName()
Implements the getName function for the Node Interface.
|
OperatorType |
getType()
Return the type of the specific operator among the
types in OperatorType.
|
protected Collection<Future<?>> |
initializeOp(org.apache.hadoop.conf.Configuration hconf)
Operator specific initialization.
|
void |
process(Object row,
int tag)
Process the row.
|
void |
setTag(byte tag) |
acceptLimitPushdown, allInitializedParentsAreClosed, areAllParentsInitialized, augmentPlan, cleanUpInputFileChanged, cleanUpInputFileChangedOp, clone, cloneOp, cloneRecursiveChildren, close, columnNamesRowResolvedCanBeObtained, completeInitializationOp, createDummy, defaultEndGroup, defaultStartGroup, dump, dump, endGroup, flush, forward, getAdditionalCounters, getChildOperators, getChildren, getColumnExprMap, getConf, getConfiguration, getDone, getExecContext, getGroupKeyObject, getIdentifier, getInputObjInspectors, getNextCntr, getNumChild, getNumParent, getOperatorId, getOperatorName, getOpTraits, getOutputObjInspector, getParentOperators, getSchema, getStatistics, getStats, initEvaluators, initEvaluators, initEvaluatorsAndReturnStruct, initialize, initialize, initializeChildren, initializeLocalWork, initOperatorId, isUseBucketizedHiveInputFormat, jobClose, jobCloseOp, logStats, opAllowedAfterMapJoin, opAllowedBeforeMapJoin, opAllowedBeforeSortMergeJoin, opAllowedConvertMapJoin, passExecContext, preorderMap, processGroup, removeChild, removeChildAndAdoptItsChildren, removeChildren, removeParent, replaceChild, replaceParent, reset, resetId, resetStats, setAlias, setChildOperators, setColumnExprMap, setConf, setDone, setExecContext, setGroupKeyObject, setId, setInputContext, setInputObjInspectors, setOperatorId, setOpTraits, setOutputCollector, setParentOperators, setReporter, setSchema, setStatistics, setUseBucketizedHiveInputFormat, startGroup, supportAutomaticSortMergeJoin, supportSkewJoinOptimization, supportUnionRemoveOptimization, toString, toString
protected Collection<Future<?>> initializeOp(org.apache.hadoop.conf.Configuration hconf) throws HiveException
Operator
initializeOp
in class Operator<SparkHashTableSinkDesc>
HiveException
public void process(Object row, int tag) throws HiveException
Operator
process
in class Operator<SparkHashTableSinkDesc>
row
- The object representing the row.tag
- The tag of the row usually means which parent this row comes from.
Rows with the same tag should have exactly the same rowInspector
all the time.HiveException
public void closeOp(boolean abort) throws HiveException
Operator
closeOp
in class Operator<SparkHashTableSinkDesc>
HiveException
protected void flushToFile(MapJoinPersistableTableContainer tableContainer, byte tag) throws Exception
Exception
public void setTag(byte tag)
public String getName()
getName
in interface Node
getName
in class Operator<SparkHashTableSinkDesc>
public OperatorType getType()
Operator
getType
in class Operator<SparkHashTableSinkDesc>
Copyright © 2017 The Apache Software Foundation. All rights reserved.