PREHOOK: query: CREATE TABLE dest1(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@dest1 POSTHOOK: query: CREATE TABLE dest1(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@dest1 PREHOOK: query: CREATE TABLE dest2(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@dest2 POSTHOOK: query: CREATE TABLE dest2(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@dest2 PREHOOK: query: -- SORT_QUERY_RESULTS EXPLAIN EXTENDED FROM srcpart INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' PREHOOK: type: QUERY POSTHOOK: query: -- SORT_QUERY_RESULTS EXPLAIN EXTENDED FROM srcpart INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' POSTHOOK: type: QUERY ABSTRACT SYNTAX TREE: TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME srcpart TOK_INSERT TOK_DESTINATION TOK_TAB TOK_TABNAME dest1 TOK_SELECT TOK_SELEXPR . TOK_TABLE_OR_COL srcpart key TOK_SELEXPR . TOK_TABLE_OR_COL srcpart value TOK_SELEXPR . TOK_TABLE_OR_COL srcpart hr TOK_SELEXPR . TOK_TABLE_OR_COL srcpart ds TOK_WHERE and and < . TOK_TABLE_OR_COL srcpart key 100 = . TOK_TABLE_OR_COL srcpart ds '2008-04-08' = . TOK_TABLE_OR_COL srcpart hr '12' TOK_INSERT TOK_DESTINATION TOK_TAB TOK_TABNAME dest2 TOK_SELECT TOK_SELEXPR . TOK_TABLE_OR_COL srcpart key TOK_SELEXPR . TOK_TABLE_OR_COL srcpart value TOK_SELEXPR . TOK_TABLE_OR_COL srcpart hr TOK_SELEXPR . TOK_TABLE_OR_COL srcpart ds TOK_WHERE and and < . TOK_TABLE_OR_COL srcpart key 100 = . TOK_TABLE_OR_COL srcpart ds '2008-04-09' = . TOK_TABLE_OR_COL srcpart hr '12' STAGE DEPENDENCIES: Stage-2 is a root stage Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 Stage-5 Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 Stage-3 depends on stages: Stage-0 Stage-4 Stage-6 Stage-7 depends on stages: Stage-6 Stage-14 depends on stages: Stage-2 , consists of Stage-11, Stage-10, Stage-12 Stage-11 Stage-1 depends on stages: Stage-11, Stage-10, Stage-13 Stage-9 depends on stages: Stage-1 Stage-10 Stage-12 Stage-13 depends on stages: Stage-12 STAGE PLANS: Stage: Stage-2 Map Reduce Map Operator Tree: TableScan alias: srcpart Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((key < 100) and (ds = '2008-04-08')) (type: boolean) Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 1 #### A masked pattern was here #### NumFilesPerFileSink: 1 Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 TotalFiles: 1 GatherStats: true MultiFileSpray: false Filter Operator isSamplingPred: false predicate: ((key < 100) and (ds = '2008-04-09')) (type: boolean) Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-09' (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 2 #### A masked pattern was here #### NumFilesPerFileSink: 1 Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 TotalFiles: 1 GatherStats: true MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: #### A masked pattern was here #### Partition base file name: hr=12 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: ds 2008-04-08 hr 12 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart numFiles 1 numRows 500 partition_columns ds/hr partition_columns.types string:string rawDataSize 5312 serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart partition_columns ds/hr partition_columns.types string:string serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart #### A masked pattern was here #### Partition base file name: hr=12 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: ds 2008-04-09 hr 12 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns key,value columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart numFiles 1 numRows 500 partition_columns ds/hr partition_columns.types string:string rawDataSize 5312 serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart partition_columns ds/hr partition_columns.types string:string serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart Truncated Path -> Alias: /srcpart/ds=2008-04-08/hr=12 [srcpart] /srcpart/ds=2008-04-09/hr=12 [srcpart] Stage: Stage-8 Conditional Operator Stage: Stage-5 Move Operator files: hdfs directory: true #### A masked pattern was here #### Stage: Stage-0 Move Operator tables: replace: true #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 Stage: Stage-3 Stats-Aggr Operator #### A masked pattern was here #### Stage: Stage-4 Map Reduce Map Operator Tree: TableScan GatherStats: false File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 TotalFiles: 1 GatherStats: false MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: #### A masked pattern was here #### Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 name: default.dest1 Truncated Path -> Alias: #### A masked pattern was here #### Stage: Stage-6 Map Reduce Map Operator Tree: TableScan GatherStats: false File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 TotalFiles: 1 GatherStats: false MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: #### A masked pattern was here #### Partition base file name: -ext-10002 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest1 serialization.ddl struct dest1 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 name: default.dest1 Truncated Path -> Alias: #### A masked pattern was here #### Stage: Stage-7 Move Operator files: hdfs directory: true #### A masked pattern was here #### Stage: Stage-14 Conditional Operator Stage: Stage-11 Move Operator files: hdfs directory: true #### A masked pattern was here #### Stage: Stage-1 Move Operator tables: replace: true #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 Stage: Stage-9 Stats-Aggr Operator #### A masked pattern was here #### Stage: Stage-10 Map Reduce Map Operator Tree: TableScan GatherStats: false File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 TotalFiles: 1 GatherStats: false MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: #### A masked pattern was here #### Partition base file name: -ext-10003 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 name: default.dest2 Truncated Path -> Alias: #### A masked pattern was here #### Stage: Stage-12 Map Reduce Map Operator Tree: TableScan GatherStats: false File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 TotalFiles: 1 GatherStats: false MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: #### A masked pattern was here #### Partition base file name: -ext-10003 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns key,value,hr,ds columns.comments columns.types int:string:string:string #### A masked pattern was here #### name default.dest2 serialization.ddl struct dest2 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2 name: default.dest2 Truncated Path -> Alias: #### A masked pattern was here #### Stage: Stage-13 Move Operator files: hdfs directory: true #### A masked pattern was here #### PREHOOK: query: FROM srcpart INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 PREHOOK: Output: default@dest1 PREHOOK: Output: default@dest2 POSTHOOK: query: FROM srcpart INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' POSTHOOK: type: QUERY POSTHOOK: Input: default@srcpart POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 POSTHOOK: Output: default@dest1 POSTHOOK: Output: default@dest2 POSTHOOK: Lineage: dest1.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] POSTHOOK: Lineage: dest1.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: dest1.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: dest2.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] POSTHOOK: Lineage: dest2.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] POSTHOOK: Lineage: dest2.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: dest2.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: SELECT dest1.* FROM dest1 sort by key,value,ds,hr PREHOOK: type: QUERY PREHOOK: Input: default@dest1 #### A masked pattern was here #### POSTHOOK: query: SELECT dest1.* FROM dest1 sort by key,value,ds,hr POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 #### A masked pattern was here #### 0 val_0 12 2008-04-08 0 val_0 12 2008-04-08 0 val_0 12 2008-04-08 10 val_10 12 2008-04-08 11 val_11 12 2008-04-08 12 val_12 12 2008-04-08 12 val_12 12 2008-04-08 15 val_15 12 2008-04-08 15 val_15 12 2008-04-08 17 val_17 12 2008-04-08 18 val_18 12 2008-04-08 18 val_18 12 2008-04-08 19 val_19 12 2008-04-08 2 val_2 12 2008-04-08 20 val_20 12 2008-04-08 24 val_24 12 2008-04-08 24 val_24 12 2008-04-08 26 val_26 12 2008-04-08 26 val_26 12 2008-04-08 27 val_27 12 2008-04-08 28 val_28 12 2008-04-08 30 val_30 12 2008-04-08 33 val_33 12 2008-04-08 34 val_34 12 2008-04-08 35 val_35 12 2008-04-08 35 val_35 12 2008-04-08 35 val_35 12 2008-04-08 37 val_37 12 2008-04-08 37 val_37 12 2008-04-08 4 val_4 12 2008-04-08 41 val_41 12 2008-04-08 42 val_42 12 2008-04-08 42 val_42 12 2008-04-08 43 val_43 12 2008-04-08 44 val_44 12 2008-04-08 47 val_47 12 2008-04-08 5 val_5 12 2008-04-08 5 val_5 12 2008-04-08 5 val_5 12 2008-04-08 51 val_51 12 2008-04-08 51 val_51 12 2008-04-08 53 val_53 12 2008-04-08 54 val_54 12 2008-04-08 57 val_57 12 2008-04-08 58 val_58 12 2008-04-08 58 val_58 12 2008-04-08 64 val_64 12 2008-04-08 65 val_65 12 2008-04-08 66 val_66 12 2008-04-08 67 val_67 12 2008-04-08 67 val_67 12 2008-04-08 69 val_69 12 2008-04-08 70 val_70 12 2008-04-08 70 val_70 12 2008-04-08 70 val_70 12 2008-04-08 72 val_72 12 2008-04-08 72 val_72 12 2008-04-08 74 val_74 12 2008-04-08 76 val_76 12 2008-04-08 76 val_76 12 2008-04-08 77 val_77 12 2008-04-08 78 val_78 12 2008-04-08 8 val_8 12 2008-04-08 80 val_80 12 2008-04-08 82 val_82 12 2008-04-08 83 val_83 12 2008-04-08 83 val_83 12 2008-04-08 84 val_84 12 2008-04-08 84 val_84 12 2008-04-08 85 val_85 12 2008-04-08 86 val_86 12 2008-04-08 87 val_87 12 2008-04-08 9 val_9 12 2008-04-08 90 val_90 12 2008-04-08 90 val_90 12 2008-04-08 90 val_90 12 2008-04-08 92 val_92 12 2008-04-08 95 val_95 12 2008-04-08 95 val_95 12 2008-04-08 96 val_96 12 2008-04-08 97 val_97 12 2008-04-08 97 val_97 12 2008-04-08 98 val_98 12 2008-04-08 98 val_98 12 2008-04-08 PREHOOK: query: SELECT dest2.* FROM dest2 sort by key,value,ds,hr PREHOOK: type: QUERY PREHOOK: Input: default@dest2 #### A masked pattern was here #### POSTHOOK: query: SELECT dest2.* FROM dest2 sort by key,value,ds,hr POSTHOOK: type: QUERY POSTHOOK: Input: default@dest2 #### A masked pattern was here #### 0 val_0 12 2008-04-09 0 val_0 12 2008-04-09 0 val_0 12 2008-04-09 10 val_10 12 2008-04-09 11 val_11 12 2008-04-09 12 val_12 12 2008-04-09 12 val_12 12 2008-04-09 15 val_15 12 2008-04-09 15 val_15 12 2008-04-09 17 val_17 12 2008-04-09 18 val_18 12 2008-04-09 18 val_18 12 2008-04-09 19 val_19 12 2008-04-09 2 val_2 12 2008-04-09 20 val_20 12 2008-04-09 24 val_24 12 2008-04-09 24 val_24 12 2008-04-09 26 val_26 12 2008-04-09 26 val_26 12 2008-04-09 27 val_27 12 2008-04-09 28 val_28 12 2008-04-09 30 val_30 12 2008-04-09 33 val_33 12 2008-04-09 34 val_34 12 2008-04-09 35 val_35 12 2008-04-09 35 val_35 12 2008-04-09 35 val_35 12 2008-04-09 37 val_37 12 2008-04-09 37 val_37 12 2008-04-09 4 val_4 12 2008-04-09 41 val_41 12 2008-04-09 42 val_42 12 2008-04-09 42 val_42 12 2008-04-09 43 val_43 12 2008-04-09 44 val_44 12 2008-04-09 47 val_47 12 2008-04-09 5 val_5 12 2008-04-09 5 val_5 12 2008-04-09 5 val_5 12 2008-04-09 51 val_51 12 2008-04-09 51 val_51 12 2008-04-09 53 val_53 12 2008-04-09 54 val_54 12 2008-04-09 57 val_57 12 2008-04-09 58 val_58 12 2008-04-09 58 val_58 12 2008-04-09 64 val_64 12 2008-04-09 65 val_65 12 2008-04-09 66 val_66 12 2008-04-09 67 val_67 12 2008-04-09 67 val_67 12 2008-04-09 69 val_69 12 2008-04-09 70 val_70 12 2008-04-09 70 val_70 12 2008-04-09 70 val_70 12 2008-04-09 72 val_72 12 2008-04-09 72 val_72 12 2008-04-09 74 val_74 12 2008-04-09 76 val_76 12 2008-04-09 76 val_76 12 2008-04-09 77 val_77 12 2008-04-09 78 val_78 12 2008-04-09 8 val_8 12 2008-04-09 80 val_80 12 2008-04-09 82 val_82 12 2008-04-09 83 val_83 12 2008-04-09 83 val_83 12 2008-04-09 84 val_84 12 2008-04-09 84 val_84 12 2008-04-09 85 val_85 12 2008-04-09 86 val_86 12 2008-04-09 87 val_87 12 2008-04-09 9 val_9 12 2008-04-09 90 val_90 12 2008-04-09 90 val_90 12 2008-04-09 90 val_90 12 2008-04-09 92 val_92 12 2008-04-09 95 val_95 12 2008-04-09 95 val_95 12 2008-04-09 96 val_96 12 2008-04-09 97 val_97 12 2008-04-09 97 val_97 12 2008-04-09 98 val_98 12 2008-04-09 98 val_98 12 2008-04-09