PREHOOK: query: explain create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 PREHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: query: explain create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 POSTHOOK: type: CREATETABLE_AS_SELECT STAGE DEPENDENCIES: Stage-1 is a root stage Stage-5 depends on stages: Stage-1 , consists of Stage-6, Stage-3 Stage-6 Stage-4 depends on stages: Stage-6 Stage-3 depends on stages: Stage-4 Stage-0 depends on stages: Stage-3 Stage-7 depends on stages: Stage-0 Stage-2 depends on stages: Stage-7 STAGE PLANS: Stage: Stage-1 Spark Edges: Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Map 4 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: TableScan alias: a Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: key (type: string) sort order: + Map-reduce partition columns: key (type: string) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE value expressions: value (type: string) Map 4 Map Operator Tree: TableScan alias: b Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: key (type: string) sort order: + Map-reduce partition columns: key (type: string) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Join Operator condition map: Inner Join 0 to 1 handleSkewJoin: true keys: 0 key (type: string) 1 key (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Stage: Stage-5 Conditional Operator Stage: Stage-6 Spark #### A masked pattern was here #### Vertices: Map 7 Map Operator Tree: TableScan Spark HashTable Sink Operator keys: 0 reducesinkkey0 (type: string) 1 reducesinkkey0 (type: string) Local Work: Map Reduce Local Work Stage: Stage-4 Spark #### A masked pattern was here #### Vertices: Map 6 Map Operator Tree: TableScan Map Join Operator condition map: Inner Join 0 to 1 keys: 0 reducesinkkey0 (type: string) 1 reducesinkkey0 (type: string) outputColumnNames: _col0, _col1 File Output Operator compressed: false table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Local Work: Map Reduce Local Work Stage: Stage-3 Spark Edges: Reducer 3 <- Map 5 (SORT, 1) #### A masked pattern was here #### Vertices: Map 5 Map Operator Tree: TableScan Reduce Output Operator key expressions: _col0 (type: string) sort order: + Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string) Reducer 3 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 30 Statistics: Num rows: 30 Data size: 300 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false Statistics: Num rows: 30 Data size: 300 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.noskew Stage: Stage-0 Move Operator files: hdfs directory: true #### A masked pattern was here #### Stage: Stage-7 Create Table Operator: Create Table columns: key string, value string input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.noskew Stage: Stage-2 Stats-Aggr Operator PREHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src PREHOOK: Output: database:default PREHOOK: Output: default@noskew POSTHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src POSTHOOK: Output: database:default POSTHOOK: Output: default@noskew PREHOOK: query: select * from noskew PREHOOK: type: QUERY PREHOOK: Input: default@noskew #### A masked pattern was here #### POSTHOOK: query: select * from noskew POSTHOOK: type: QUERY POSTHOOK: Input: default@noskew #### A masked pattern was here #### 0 val_0 0 val_0 0 val_0 0 val_0 0 val_0 0 val_0 0 val_0 0 val_0 0 val_0 10 val_10 100 val_100 100 val_100 100 val_100 100 val_100 103 val_103 103 val_103 103 val_103 103 val_103 104 val_104 104 val_104 104 val_104 104 val_104 105 val_105 11 val_11 111 val_111 113 val_113 113 val_113 113 val_113 113 val_113 114 val_114