PREHOOK: query: drop table dest1 PREHOOK: type: DROPTABLE POSTHOOK: query: drop table dest1 POSTHOOK: type: DROPTABLE PREHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default POSTHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@dest1 PREHOOK: query: EXPLAIN FROM ( FROM src SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter' USING 'cat' AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader' WHERE key < 100 ) tmap INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue PREHOOK: type: QUERY POSTHOOK: query: EXPLAIN FROM ( FROM src SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter' USING 'cat' AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader' WHERE key < 100 ) tmap INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 Map Reduce Map Operator Tree: TableScan alias: src Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (key < 100) (type: boolean) Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: UDFToByte(key) (type: tinyint), value (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE Transform Operator command: cat output info: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string) Reduce Operator Tree: Extract Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false Statistics: Num rows: 9 Data size: 1803 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 Stage: Stage-0 Move Operator tables: replace: true table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 Stage: Stage-2 Stats-Aggr Operator PREHOOK: query: FROM ( FROM src SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter' USING 'cat' AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader' WHERE key < 100 ) tmap INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@dest1 POSTHOOK: query: FROM ( FROM src SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter' USING 'cat' AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe' RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader' WHERE key < 100 ) tmap INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: default@dest1 POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 #### A masked pattern was here #### POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 #### A masked pattern was here #### POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] 0 val_0 0 val_0 0 val_0 10 val_10 11 val_11 12 val_12 12 val_12 15 val_15 15 val_15 17 val_17 18 val_18 18 val_18 19 val_19 2 val_2 20 val_20 24 val_24 24 val_24 26 val_26 26 val_26 27 val_27 28 val_28 30 val_30 33 val_33 34 val_34 35 val_35 35 val_35 35 val_35 37 val_37 37 val_37 4 val_4 41 val_41 42 val_42 42 val_42 43 val_43 44 val_44 47 val_47 5 val_5 5 val_5 5 val_5 51 val_51 51 val_51 53 val_53 54 val_54 57 val_57 58 val_58 58 val_58 64 val_64 65 val_65 66 val_66 67 val_67 67 val_67 69 val_69 70 val_70 70 val_70 70 val_70 72 val_72 72 val_72 74 val_74 76 val_76 76 val_76 77 val_77 78 val_78 8 val_8 80 val_80 82 val_82 83 val_83 83 val_83 84 val_84 84 val_84 85 val_85 86 val_86 87 val_87 9 val_9 90 val_90 90 val_90 90 val_90 92 val_92 95 val_95 95 val_95 96 val_96 97 val_97 97 val_97 98 val_98 98 val_98 PREHOOK: query: drop table dest1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@dest1 PREHOOK: Output: default@dest1 POSTHOOK: query: drop table dest1 POSTHOOK: type: DROPTABLE POSTHOOK: Input: default@dest1 POSTHOOK: Output: default@dest1 POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]