| PREHOOK: query: create table inputTbl1_n7(key string, val string) stored as textfile |
| PREHOOK: type: CREATETABLE |
| PREHOOK: Output: database:default |
| PREHOOK: Output: default@inputTbl1_n7 |
| POSTHOOK: query: create table inputTbl1_n7(key string, val string) stored as textfile |
| POSTHOOK: type: CREATETABLE |
| POSTHOOK: Output: database:default |
| POSTHOOK: Output: default@inputTbl1_n7 |
| PREHOOK: query: create table outputTbl1_n9(key string, `values` bigint) stored as rcfile |
| PREHOOK: type: CREATETABLE |
| PREHOOK: Output: database:default |
| PREHOOK: Output: default@outputTbl1_n9 |
| POSTHOOK: query: create table outputTbl1_n9(key string, `values` bigint) stored as rcfile |
| POSTHOOK: type: CREATETABLE |
| POSTHOOK: Output: database:default |
| POSTHOOK: Output: default@outputTbl1_n9 |
| PREHOOK: query: load data local inpath '../../data/files/T1.txt' into table inputTbl1_n7 |
| PREHOOK: type: LOAD |
| #### A masked pattern was here #### |
| PREHOOK: Output: default@inputtbl1_n7 |
| POSTHOOK: query: load data local inpath '../../data/files/T1.txt' into table inputTbl1_n7 |
| POSTHOOK: type: LOAD |
| #### A masked pattern was here #### |
| POSTHOOK: Output: default@inputtbl1_n7 |
| PREHOOK: query: explain |
| insert overwrite table outputTbl1_n9 |
| SELECT * FROM |
| ( |
| select key, 1 as `values` from inputTbl1_n7 |
| union all |
| select * FROM ( |
| SELECT key, count(1) as `values` from inputTbl1_n7 group by key |
| UNION ALL |
| SELECT key, 2 as `values` from inputTbl1_n7 |
| ) a |
| )b |
| PREHOOK: type: QUERY |
| POSTHOOK: query: explain |
| insert overwrite table outputTbl1_n9 |
| SELECT * FROM |
| ( |
| select key, 1 as `values` from inputTbl1_n7 |
| union all |
| select * FROM ( |
| SELECT key, count(1) as `values` from inputTbl1_n7 group by key |
| UNION ALL |
| SELECT key, 2 as `values` from inputTbl1_n7 |
| ) a |
| )b |
| POSTHOOK: type: QUERY |
| STAGE DEPENDENCIES: |
| Stage-1 is a root stage |
| Stage-6 depends on stages: Stage-1, Stage-7, Stage-8 , consists of Stage-3, Stage-2, Stage-4 |
| Stage-3 |
| Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 |
| Stage-2 |
| Stage-4 |
| Stage-5 depends on stages: Stage-4 |
| Stage-7 is a root stage |
| Stage-8 is a root stage |
| |
| STAGE PLANS: |
| Stage: Stage-1 |
| Map Reduce |
| Map Operator Tree: |
| TableScan |
| alias: inputtbl1_n7 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| Select Operator |
| expressions: key (type: string), 1L (type: bigint) |
| outputColumnNames: _col0, _col1 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| table: |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe |
| name: default.outputtbl1_n9 |
| Execution mode: vectorized |
| |
| Stage: Stage-6 |
| Conditional Operator |
| |
| Stage: Stage-3 |
| Move Operator |
| files: |
| hdfs directory: true |
| #### A masked pattern was here #### |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: true |
| table: |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe |
| name: default.outputtbl1_n9 |
| |
| Stage: Stage-2 |
| Merge File Operator |
| Map Operator Tree: |
| RCFile Merge Operator |
| merge level: block |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| |
| Stage: Stage-4 |
| Merge File Operator |
| Map Operator Tree: |
| RCFile Merge Operator |
| merge level: block |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| |
| Stage: Stage-5 |
| Move Operator |
| files: |
| hdfs directory: true |
| #### A masked pattern was here #### |
| |
| Stage: Stage-7 |
| Map Reduce |
| Map Operator Tree: |
| TableScan |
| alias: inputtbl1_n7 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| Select Operator |
| expressions: key (type: string) |
| outputColumnNames: key |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| Group By Operator |
| aggregations: count() |
| keys: key (type: string) |
| mode: hash |
| outputColumnNames: _col0, _col1 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| Reduce Output Operator |
| key expressions: _col0 (type: string) |
| sort order: + |
| Map-reduce partition columns: _col0 (type: string) |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| value expressions: _col1 (type: bigint) |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Group By Operator |
| aggregations: count(VALUE._col0) |
| keys: KEY._col0 (type: string) |
| mode: mergepartial |
| outputColumnNames: _col0, _col1 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| table: |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe |
| name: default.outputtbl1_n9 |
| |
| Stage: Stage-8 |
| Map Reduce |
| Map Operator Tree: |
| TableScan |
| alias: inputtbl1_n7 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| Select Operator |
| expressions: key (type: string), 2L (type: bigint) |
| outputColumnNames: _col0, _col1 |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE |
| table: |
| input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe |
| name: default.outputtbl1_n9 |
| Execution mode: vectorized |
| |
| PREHOOK: query: insert overwrite table outputTbl1_n9 |
| SELECT * FROM |
| ( |
| select key, 1 as `values` from inputTbl1_n7 |
| union all |
| select * FROM ( |
| SELECT key, count(1) as `values` from inputTbl1_n7 group by key |
| UNION ALL |
| SELECT key, 2 as `values` from inputTbl1_n7 |
| ) a |
| )b |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@inputtbl1_n7 |
| PREHOOK: Output: default@outputtbl1_n9 |
| POSTHOOK: query: insert overwrite table outputTbl1_n9 |
| SELECT * FROM |
| ( |
| select key, 1 as `values` from inputTbl1_n7 |
| union all |
| select * FROM ( |
| SELECT key, count(1) as `values` from inputTbl1_n7 group by key |
| UNION ALL |
| SELECT key, 2 as `values` from inputTbl1_n7 |
| ) a |
| )b |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@inputtbl1_n7 |
| POSTHOOK: Output: default@outputtbl1_n9 |
| POSTHOOK: Lineage: outputtbl1_n9.key EXPRESSION [(inputtbl1_n7)inputtbl1_n7.FieldSchema(name:key, type:string, comment:null), ] |
| POSTHOOK: Lineage: outputtbl1_n9.values EXPRESSION [(inputtbl1_n7)inputtbl1_n7.null, ] |
| PREHOOK: query: desc formatted outputTbl1_n9 |
| PREHOOK: type: DESCTABLE |
| PREHOOK: Input: default@outputtbl1_n9 |
| POSTHOOK: query: desc formatted outputTbl1_n9 |
| POSTHOOK: type: DESCTABLE |
| POSTHOOK: Input: default@outputtbl1_n9 |
| # col_name data_type comment |
| key string |
| values bigint |
| |
| # Detailed Table Information |
| Database: default |
| #### A masked pattern was here #### |
| Retention: 0 |
| #### A masked pattern was here #### |
| Table Type: MANAGED_TABLE |
| Table Parameters: |
| bucketing_version 2 |
| numFiles 3 |
| totalSize 271 |
| #### A masked pattern was here #### |
| |
| # Storage Information |
| SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe |
| InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat |
| OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat |
| Compressed: No |
| Num Buckets: -1 |
| Bucket Columns: [] |
| Sort Columns: [] |
| Storage Desc Params: |
| serialization.format 1 |
| PREHOOK: query: select * from outputTbl1_n9 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@outputtbl1_n9 |
| #### A masked pattern was here #### |
| POSTHOOK: query: select * from outputTbl1_n9 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@outputtbl1_n9 |
| #### A masked pattern was here #### |
| 1 1 |
| 1 1 |
| 1 2 |
| 2 1 |
| 2 1 |
| 2 2 |
| 3 1 |
| 3 1 |
| 3 2 |
| 7 1 |
| 7 1 |
| 7 2 |
| 8 1 |
| 8 1 |
| 8 2 |
| 8 2 |
| 8 2 |