PREHOOK: query: CREATE TABLE T1(key STRING, val STRING) PARTITIONED BY (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@T1 POSTHOOK: query: CREATE TABLE T1(key STRING, val STRING) PARTITIONED BY (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@T1 PREHOOK: query: -- perform an insert to make sure there are 2 files INSERT OVERWRITE TABLE T1 PARTITION (ds='1') SELECT * from src where key = 0 or key = 11 PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@t1@ds=1 POSTHOOK: query: -- perform an insert to make sure there are 2 files INSERT OVERWRITE TABLE T1 PARTITION (ds='1') SELECT * from src where key = 0 or key = 11 POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: default@t1@ds=1 POSTHOOK: Lineage: t1 PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t1 PARTITION(ds=1).val SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: -- The plan is converted to a map-side plan EXPLAIN select distinct key from T1 PREHOOK: type: QUERY POSTHOOK: query: -- The plan is converted to a map-side plan EXPLAIN select distinct key from T1 POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 Map Reduce Map Operator Tree: TableScan alias: t1 Statistics: Num rows: 4 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 Statistics: Num rows: 4 Data size: 30 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: string) mode: final outputColumnNames: _col0 Statistics: Num rows: 2 Data size: 15 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 Statistics: Num rows: 2 Data size: 15 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false Statistics: Num rows: 2 Data size: 15 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: ListSink PREHOOK: query: select distinct key from T1 PREHOOK: type: QUERY PREHOOK: Input: default@t1 PREHOOK: Input: default@t1@ds=1 #### A masked pattern was here #### POSTHOOK: query: select distinct key from T1 POSTHOOK: type: QUERY POSTHOOK: Input: default@t1 POSTHOOK: Input: default@t1@ds=1 #### A masked pattern was here #### 0 11 PREHOOK: query: -- perform an insert to make sure there are 2 files INSERT OVERWRITE TABLE T1 PARTITION (ds='2') SELECT * from src where key = 0 or key = 11 PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@t1@ds=2 POSTHOOK: query: -- perform an insert to make sure there are 2 files INSERT OVERWRITE TABLE T1 PARTITION (ds='2') SELECT * from src where key = 0 or key = 11 POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: default@t1@ds=2 POSTHOOK: Lineage: t1 PARTITION(ds=2).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t1 PARTITION(ds=2).val SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: -- The plan is not converted to a map-side, since although the sorting columns and grouping -- columns match, the user is querying multiple input partitions EXPLAIN select distinct key from T1 PREHOOK: type: QUERY POSTHOOK: query: -- The plan is not converted to a map-side, since although the sorting columns and grouping -- columns match, the user is querying multiple input partitions EXPLAIN select distinct key from T1 POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 Map Reduce Map Operator Tree: TableScan alias: t1 Statistics: Num rows: 8 Data size: 60 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 Statistics: Num rows: 8 Data size: 60 Basic stats: COMPLETE Column stats: NONE Group By Operator bucketGroup: true keys: _col0 (type: string) mode: hash outputColumnNames: _col0 Statistics: Num rows: 8 Data size: 60 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 8 Data size: 60 Basic stats: COMPLETE Column stats: NONE Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 4 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 Statistics: Num rows: 4 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false Statistics: Num rows: 4 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: ListSink PREHOOK: query: select distinct key from T1 PREHOOK: type: QUERY PREHOOK: Input: default@t1 PREHOOK: Input: default@t1@ds=1 PREHOOK: Input: default@t1@ds=2 #### A masked pattern was here #### POSTHOOK: query: select distinct key from T1 POSTHOOK: type: QUERY POSTHOOK: Input: default@t1 POSTHOOK: Input: default@t1@ds=1 POSTHOOK: Input: default@t1@ds=2 #### A masked pattern was here #### 0 11 PREHOOK: query: DROP TABLE T1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@t1 PREHOOK: Output: default@t1 POSTHOOK: query: DROP TABLE T1 POSTHOOK: type: DROPTABLE POSTHOOK: Input: default@t1 POSTHOOK: Output: default@t1