| PREHOOK: query: create external table ice01(a int, b string, c int) stored by iceberg stored as orc tblproperties ('format-version'='2') |
| PREHOOK: type: CREATETABLE |
| PREHOOK: Output: database:default |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: create external table ice01(a int, b string, c int) stored by iceberg stored as orc tblproperties ('format-version'='2') |
| POSTHOOK: type: CREATETABLE |
| POSTHOOK: Output: database:default |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: create table source01(a int, b string, c int) |
| PREHOOK: type: CREATETABLE |
| PREHOOK: Output: database:default |
| PREHOOK: Output: default@source01 |
| POSTHOOK: query: create table source01(a int, b string, c int) |
| POSTHOOK: type: CREATETABLE |
| POSTHOOK: Output: database:default |
| POSTHOOK: Output: default@source01 |
| PREHOOK: query: insert into ice01 values (1, 'one', 50), (2, 'two', 51), (111, 'one', 55) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: insert into ice01 values (1, 'one', 50), (2, 'two', 51), (111, 'one', 55) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: insert into source01 values (1, 'one', 50), (2, 'two', 51), (3, 'three', 52), (4, 'four', 53), (5, 'five', 54), (111, 'one', 55) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@source01 |
| POSTHOOK: query: insert into source01 values (1, 'one', 50), (2, 'two', 51), (3, 'three', 52), (4, 'four', 53), (5, 'five', 54), (111, 'one', 55) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@source01 |
| POSTHOOK: Lineage: source01.a SCRIPT [] |
| POSTHOOK: Lineage: source01.b SCRIPT [] |
| POSTHOOK: Lineage: source01.c SCRIPT [] |
| PREHOOK: query: alter table ice01 create branch test1 |
| PREHOOK: type: ALTERTABLE_CREATEBRANCH |
| PREHOOK: Input: default@ice01 |
| POSTHOOK: query: alter table ice01 create branch test1 |
| POSTHOOK: type: ALTERTABLE_CREATEBRANCH |
| POSTHOOK: Input: default@ice01 |
| PREHOOK: query: explain select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: explain select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| STAGE DEPENDENCIES: |
| Stage-0 is a root stage |
| |
| STAGE PLANS: |
| Stage: Stage-0 |
| Fetch Operator |
| limit: -1 |
| Processor Tree: |
| TableScan |
| alias: ice01 |
| branch name: branch_test1 |
| Select Operator |
| expressions: a (type: int), b (type: string), c (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| ListSink |
| |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 one 50 |
| 111 one 55 |
| 2 two 51 |
| PREHOOK: query: select * from ice01 for system_version as of 'test1' |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from ice01 for system_version as of 'test1' |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 one 50 |
| 111 one 55 |
| 2 two 51 |
| PREHOOK: query: explain insert into default.ice01.branch_test1 values(22, 'three', 44), (33, 'three', 66) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: explain insert into default.ice01.branch_test1 values(22, 'three', 44), (33, 'three', 66) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@ice01 |
| STAGE DEPENDENCIES: |
| Stage-1 is a root stage |
| Stage-2 depends on stages: Stage-1 |
| Stage-0 depends on stages: Stage-2 |
| Stage-3 depends on stages: Stage-0 |
| |
| STAGE PLANS: |
| Stage: Stage-1 |
| Tez |
| #### A masked pattern was here #### |
| Edges: |
| Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: _dummy_table |
| Row Limit Per Split: 1 |
| Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: array(const struct(22,'three',44),const struct(33,'three',66)) (type: array<struct<col1:int,col2:string,col3:int>>) |
| outputColumnNames: _col0 |
| Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE Column stats: COMPLETE |
| UDTF Operator |
| Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE Column stats: COMPLETE |
| function name: inline |
| Select Operator |
| expressions: col1 (type: int), col2 (type: string), col3 (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Select Operator |
| expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) |
| outputColumnNames: a, b, c |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| Group By Operator |
| aggregations: min(a), max(a), count(1), count(a), compute_bit_vector_hll(a), max(length(b)), avg(COALESCE(length(b),0)), count(b), compute_bit_vector_hll(b), min(c), max(c), count(c), compute_bit_vector_hll(c) |
| minReductionHashAggr: 0.4 |
| mode: hash |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 |
| Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| null sort order: |
| sort order: |
| Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: binary), _col5 (type: int), _col6 (type: struct<count:bigint,sum:double,input:int>), _col7 (type: bigint), _col8 (type: binary), _col9 (type: int), _col10 (type: int), _col11 (type: bigint), _col12 (type: binary) |
| Reducer 2 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Group By Operator |
| aggregations: min(VALUE._col0), max(VALUE._col1), count(VALUE._col2), count(VALUE._col3), compute_bit_vector_hll(VALUE._col4), max(VALUE._col5), avg(VALUE._col6), count(VALUE._col7), compute_bit_vector_hll(VALUE._col8), min(VALUE._col9), max(VALUE._col10), count(VALUE._col11), compute_bit_vector_hll(VALUE._col12) |
| mode: mergepartial |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 |
| Statistics: Num rows: 1 Data size: 492 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: 'LONG' (type: string), UDFToLong(_col0) (type: bigint), UDFToLong(_col1) (type: bigint), (_col2 - _col3) (type: bigint), COALESCE(ndv_compute_bit_vector(_col4),0) (type: bigint), _col4 (type: binary), 'STRING' (type: string), UDFToLong(COALESCE(_col5,0)) (type: bigint), COALESCE(_col6,0) (type: double), (_col2 - _col7) (type: bigint), COALESCE(ndv_compute_bit_vector(_col8),0) (type: bigint), _col8 (type: binary), 'LONG' (type: string), UDFToLong(_col9) (type: bigint), UDFToLong(_col10) (type: bigint), (_col2 - _col11) (type: bigint), COALESCE(ndv_compute_bit_vector(_col12),0) (type: bigint), _col12 (type: binary) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 |
| Statistics: Num rows: 1 Data size: 794 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 794 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.hadoop.mapred.SequenceFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| |
| Stage: Stage-2 |
| Dependency Collection |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: false |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-3 |
| Stats Work |
| Basic Stats Work: |
| Column Stats Desc: |
| Columns: a, b, c |
| Column Types: int, string, int |
| Table: default.ice01 |
| |
| PREHOOK: query: insert into default.ice01.branch_test1 values(22, 'three', 44), (33, 'three', 66) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: insert into default.ice01.branch_test1 values(22, 'three', 44), (33, 'three', 66) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 one 50 |
| 111 one 55 |
| 2 two 51 |
| 22 three 44 |
| 33 three 66 |
| PREHOOK: query: explain delete from default.ice01.branch_test1 where a=22 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: explain delete from default.ice01.branch_test1 where a=22 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| STAGE DEPENDENCIES: |
| Stage-1 is a root stage |
| Stage-2 depends on stages: Stage-1 |
| Stage-0 depends on stages: Stage-2 |
| Stage-3 depends on stages: Stage-0 |
| |
| STAGE PLANS: |
| Stage: Stage-1 |
| Tez |
| #### A masked pattern was here #### |
| Edges: |
| Reducer 2 <- Map 1 (SIMPLE_EDGE) |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: ice01 |
| branch name: branch_test1 |
| filterExpr: (a = 22) (type: boolean) |
| Statistics: Num rows: 3 Data size: 291 Basic stats: COMPLETE Column stats: COMPLETE |
| Filter Operator |
| predicate: (a = 22) (type: boolean) |
| Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: PARTITION__SPEC__ID (type: int), PARTITION__HASH (type: bigint), FILE__PATH (type: string), ROW__POSITION (type: bigint), b (type: string), c (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col5, _col6 |
| Statistics: Num rows: 2 Data size: 594 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| null sort order: zzzz |
| sort order: ++++ |
| Statistics: Num rows: 2 Data size: 594 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col5 (type: string), _col6 (type: int) |
| Execution mode: vectorized |
| Reducer 2 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Select Operator |
| expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: bigint), 22 (type: int), VALUE._col0 (type: string), VALUE._col1 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 2 Data size: 602 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 2 Data size: 602 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-2 |
| Dependency Collection |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: false |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-3 |
| Stats Work |
| Basic Stats Work: |
| |
| PREHOOK: query: delete from default.ice01.branch_test1 where a=22 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: delete from default.ice01.branch_test1 where a=22 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 one 50 |
| 111 one 55 |
| 2 two 51 |
| 33 three 66 |
| PREHOOK: query: explain update default.ice01.branch_test1 set a=33 where c=66 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: explain update default.ice01.branch_test1 set a=33 where c=66 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| STAGE DEPENDENCIES: |
| Stage-2 is a root stage |
| Stage-3 depends on stages: Stage-2 |
| Stage-0 depends on stages: Stage-3 |
| Stage-4 depends on stages: Stage-0 |
| |
| STAGE PLANS: |
| Stage: Stage-2 |
| Tez |
| #### A masked pattern was here #### |
| Edges: |
| Reducer 2 <- Map 1 (SIMPLE_EDGE) |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: ice01 |
| branch name: branch_test1 |
| filterExpr: (c = 66) (type: boolean) |
| Statistics: Num rows: 3 Data size: 291 Basic stats: COMPLETE Column stats: COMPLETE |
| Filter Operator |
| predicate: (c = 66) (type: boolean) |
| Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: PARTITION__SPEC__ID (type: int), PARTITION__HASH (type: bigint), FILE__PATH (type: string), ROW__POSITION (type: bigint), a (type: int), b (type: string), b (type: string) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col8 |
| Statistics: Num rows: 2 Data size: 772 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint), _col4 (type: int), _col5 (type: string) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 |
| Statistics: Num rows: 2 Data size: 594 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| null sort order: zzzz |
| sort order: ++++ |
| Statistics: Num rows: 2 Data size: 594 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col4 (type: int), _col5 (type: string) |
| Select Operator |
| expressions: 33 (type: int), _col8 (type: string), 66 (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Execution mode: vectorized |
| Reducer 2 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Select Operator |
| expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: bigint), VALUE._col0 (type: int), VALUE._col1 (type: string), 66 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 2 Data size: 602 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 2 Data size: 602 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-3 |
| Dependency Collection |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: false |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-4 |
| Stats Work |
| Basic Stats Work: |
| |
| PREHOOK: query: update default.ice01.branch_test1 set a=33 where c=66 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: update default.ice01.branch_test1 set a=33 where c=66 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 one 50 |
| 111 one 55 |
| 2 two 51 |
| 33 three 66 |
| PREHOOK: query: explain |
| merge into default.ice01.branch_test1 as t using source01 src ON t.a = src.a |
| when matched and t.a > 100 THEN DELETE |
| when matched then update set b = 'Merged', c = t.c + 10 |
| when not matched then insert values (src.a, src.b, src.c) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Input: default@source01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@merge_tmp_table |
| POSTHOOK: query: explain |
| merge into default.ice01.branch_test1 as t using source01 src ON t.a = src.a |
| when matched and t.a > 100 THEN DELETE |
| when matched then update set b = 'Merged', c = t.c + 10 |
| when not matched then insert values (src.a, src.b, src.c) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Input: default@source01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@merge_tmp_table |
| STAGE DEPENDENCIES: |
| Stage-5 is a root stage |
| Stage-6 depends on stages: Stage-5 |
| Stage-0 depends on stages: Stage-6 |
| Stage-7 depends on stages: Stage-0 |
| Stage-4 depends on stages: Stage-6 |
| Stage-8 depends on stages: Stage-4 |
| |
| STAGE PLANS: |
| Stage: Stage-5 |
| Tez |
| #### A masked pattern was here #### |
| Edges: |
| Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) |
| Reducer 3 <- Reducer 2 (SIMPLE_EDGE) |
| Reducer 4 <- Reducer 2 (SIMPLE_EDGE) |
| Reducer 5 <- Reducer 2 (SIMPLE_EDGE) |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: src |
| Statistics: Num rows: 6 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: a (type: int), b (type: string), c (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 6 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int) |
| null sort order: z |
| sort order: + |
| Map-reduce partition columns: _col0 (type: int) |
| Statistics: Num rows: 6 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col1 (type: string), _col2 (type: int) |
| Execution mode: vectorized |
| Map 6 |
| Map Operator Tree: |
| TableScan |
| alias: ice01 |
| branch name: branch_test1 |
| filterExpr: a is not null (type: boolean) |
| Statistics: Num rows: 3 Data size: 291 Basic stats: COMPLETE Column stats: COMPLETE |
| Filter Operator |
| predicate: a is not null (type: boolean) |
| Statistics: Num rows: 3 Data size: 291 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: PARTITION__SPEC__ID (type: int), PARTITION__HASH (type: bigint), FILE__PATH (type: string), ROW__POSITION (type: bigint), a (type: int), b (type: string), c (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 3 Data size: 903 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col4 (type: int) |
| null sort order: z |
| sort order: + |
| Map-reduce partition columns: _col4 (type: int) |
| Statistics: Num rows: 3 Data size: 903 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint), _col5 (type: string), _col6 (type: int) |
| Execution mode: vectorized |
| Reducer 2 |
| Reduce Operator Tree: |
| Merge Join Operator |
| condition map: |
| Left Outer Join 0 to 1 |
| keys: |
| 0 _col0 (type: int) |
| 1 _col4 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 |
| Statistics: Num rows: 3 Data size: 1191 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col1 (type: string), _col0 (type: int), _col5 (type: string), _col2 (type: int), _col6 (type: bigint), _col4 (type: bigint), _col3 (type: int), _col9 (type: int), _col8 (type: string), _col7 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 |
| Statistics: Num rows: 3 Data size: 1191 Basic stats: COMPLETE Column stats: COMPLETE |
| Filter Operator |
| predicate: ((_col9 = _col1) and (_col9 > 100)) (type: boolean) |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col6 (type: int), _col5 (type: bigint), _col2 (type: string), _col4 (type: bigint), _col9 (type: int), _col8 (type: string), _col7 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| null sort order: zzzz |
| sort order: ++++ |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col4 (type: int), _col5 (type: string), _col6 (type: int) |
| Filter Operator |
| predicate: ((_col9 = _col1) and (_col9 <= 100)) (type: boolean) |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col6 (type: int), _col5 (type: bigint), _col2 (type: string), _col4 (type: bigint), _col9 (type: int), _col8 (type: string), _col7 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| null sort order: zzzz |
| sort order: ++++ |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col4 (type: int), _col5 (type: string), _col6 (type: int) |
| Filter Operator |
| predicate: ((_col9 = _col1) and (_col9 <= 100)) (type: boolean) |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col9 (type: int), 'Merged' (type: string), (_col7 + 10) (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 1 Data size: 98 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 98 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Filter Operator |
| predicate: _col9 is null (type: boolean) |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col1 (type: int), _col0 (type: string), _col3 (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Filter Operator |
| predicate: (_col9 = _col1) (type: boolean) |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: _col2 (type: string), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: int) |
| outputColumnNames: _col2, _col4, _col5, _col6 |
| Statistics: Num rows: 1 Data size: 397 Basic stats: COMPLETE Column stats: COMPLETE |
| Group By Operator |
| aggregations: count() |
| keys: _col6 (type: int), _col5 (type: bigint), _col2 (type: string), _col4 (type: bigint) |
| minReductionHashAggr: 0.4 |
| mode: hash |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4 |
| Statistics: Num rows: 1 Data size: 212 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| key expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| null sort order: zzzz |
| sort order: ++++ |
| Map-reduce partition columns: _col0 (type: int), _col1 (type: bigint), _col2 (type: string), _col3 (type: bigint) |
| Statistics: Num rows: 1 Data size: 212 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col4 (type: bigint) |
| Reducer 3 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Select Operator |
| expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: bigint), VALUE._col0 (type: int), VALUE._col1 (type: string), VALUE._col2 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Reducer 4 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Select Operator |
| expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: bigint), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: bigint), VALUE._col0 (type: int), VALUE._col1 (type: string), VALUE._col2 (type: int) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 301 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Reducer 5 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Group By Operator |
| aggregations: count(VALUE._col0) |
| keys: KEY._col0 (type: int), KEY._col1 (type: bigint), KEY._col2 (type: string), KEY._col3 (type: bigint) |
| mode: mergepartial |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4 |
| Statistics: Num rows: 1 Data size: 212 Basic stats: COMPLETE Column stats: COMPLETE |
| Filter Operator |
| predicate: (_col4 > 1L) (type: boolean) |
| Statistics: Num rows: 1 Data size: 212 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: cardinality_violation(_col0,_col1,_col2,_col3) (type: int) |
| outputColumnNames: _col0 |
| Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.hadoop.mapred.TextInputFormat |
| output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
| serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| name: default.merge_tmp_table |
| |
| Stage: Stage-6 |
| Dependency Collection |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: false |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-7 |
| Stats Work |
| Basic Stats Work: |
| |
| Stage: Stage-4 |
| Move Operator |
| tables: |
| replace: false |
| table: |
| input format: org.apache.hadoop.mapred.TextInputFormat |
| output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
| serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| name: default.merge_tmp_table |
| |
| Stage: Stage-8 |
| Stats Work |
| Basic Stats Work: |
| |
| PREHOOK: query: merge into default.ice01.branch_test1 as t using source01 src ON t.a = src.a |
| when matched and t.a > 100 THEN DELETE |
| when matched then update set b = 'Merged', c = t.c + 10 |
| when not matched then insert values (src.a, src.b, src.c) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Input: default@source01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@ice01 |
| PREHOOK: Output: default@merge_tmp_table |
| POSTHOOK: query: merge into default.ice01.branch_test1 as t using source01 src ON t.a = src.a |
| when matched and t.a > 100 THEN DELETE |
| when matched then update set b = 'Merged', c = t.c + 10 |
| when not matched then insert values (src.a, src.b, src.c) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Input: default@source01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| POSTHOOK: Output: default@merge_tmp_table |
| POSTHOOK: Lineage: merge_tmp_table.val EXPRESSION [(ice01)ice01.null, ] |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 1 Merged 60 |
| 2 Merged 61 |
| 3 three 52 |
| 33 three 66 |
| 4 four 53 |
| 5 five 54 |
| PREHOOK: query: explain insert overwrite table default.ice01.branch_test1 values (77, 'one', 88) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: explain insert overwrite table default.ice01.branch_test1 values (77, 'one', 88) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@ice01 |
| STAGE DEPENDENCIES: |
| Stage-1 is a root stage |
| Stage-2 depends on stages: Stage-1 |
| Stage-0 depends on stages: Stage-2 |
| Stage-3 depends on stages: Stage-0 |
| |
| STAGE PLANS: |
| Stage: Stage-1 |
| Tez |
| #### A masked pattern was here #### |
| Edges: |
| Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: _dummy_table |
| Row Limit Per Split: 1 |
| Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: array(const struct(77,'one',88)) (type: array<struct<col1:int,col2:string,col3:int>>) |
| outputColumnNames: _col0 |
| Statistics: Num rows: 1 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE |
| UDTF Operator |
| Statistics: Num rows: 1 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE |
| function name: inline |
| Select Operator |
| expressions: col1 (type: int), col2 (type: string), col3 (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| Select Operator |
| expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) |
| outputColumnNames: a, b, c |
| Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE |
| Group By Operator |
| aggregations: min(a), max(a), count(1), count(a), compute_bit_vector_hll(a), max(length(b)), avg(COALESCE(length(b),0)), count(b), compute_bit_vector_hll(b), min(c), max(c), count(c), compute_bit_vector_hll(c) |
| minReductionHashAggr: 0.4 |
| mode: hash |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 |
| Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE |
| Reduce Output Operator |
| null sort order: |
| sort order: |
| Statistics: Num rows: 1 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE |
| value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: binary), _col5 (type: int), _col6 (type: struct<count:bigint,sum:double,input:int>), _col7 (type: bigint), _col8 (type: binary), _col9 (type: int), _col10 (type: int), _col11 (type: bigint), _col12 (type: binary) |
| Reducer 2 |
| Execution mode: vectorized |
| Reduce Operator Tree: |
| Group By Operator |
| aggregations: min(VALUE._col0), max(VALUE._col1), count(VALUE._col2), count(VALUE._col3), compute_bit_vector_hll(VALUE._col4), max(VALUE._col5), avg(VALUE._col6), count(VALUE._col7), compute_bit_vector_hll(VALUE._col8), min(VALUE._col9), max(VALUE._col10), count(VALUE._col11), compute_bit_vector_hll(VALUE._col12) |
| mode: mergepartial |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 |
| Statistics: Num rows: 1 Data size: 492 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: 'LONG' (type: string), UDFToLong(_col0) (type: bigint), UDFToLong(_col1) (type: bigint), (_col2 - _col3) (type: bigint), COALESCE(ndv_compute_bit_vector(_col4),0) (type: bigint), _col4 (type: binary), 'STRING' (type: string), UDFToLong(COALESCE(_col5,0)) (type: bigint), COALESCE(_col6,0) (type: double), (_col2 - _col7) (type: bigint), COALESCE(ndv_compute_bit_vector(_col8),0) (type: bigint), _col8 (type: binary), 'LONG' (type: string), UDFToLong(_col9) (type: bigint), UDFToLong(_col10) (type: bigint), (_col2 - _col11) (type: bigint), COALESCE(ndv_compute_bit_vector(_col12),0) (type: bigint), _col12 (type: binary) |
| outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 |
| Statistics: Num rows: 1 Data size: 794 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 1 Data size: 794 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.hadoop.mapred.SequenceFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| |
| Stage: Stage-2 |
| Dependency Collection |
| |
| Stage: Stage-0 |
| Move Operator |
| tables: |
| replace: true |
| table: |
| input format: org.apache.iceberg.mr.hive.HiveIcebergInputFormat |
| output format: org.apache.iceberg.mr.hive.HiveIcebergOutputFormat |
| serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe |
| name: default.ice01 |
| |
| Stage: Stage-3 |
| Stats Work |
| Basic Stats Work: |
| Column Stats Desc: |
| Columns: a, b, c |
| Column Types: int, string, int |
| Table: default.ice01 |
| |
| PREHOOK: query: insert overwrite table default.ice01.branch_test1 values (77, 'one', 88) |
| PREHOOK: type: QUERY |
| PREHOOK: Input: _dummy_database@_dummy_table |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: insert overwrite table default.ice01.branch_test1 values (77, 'one', 88) |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: _dummy_database@_dummy_table |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 77 one 88 |
| PREHOOK: query: explain select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: explain select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| STAGE DEPENDENCIES: |
| Stage-1 is a root stage |
| Stage-0 depends on stages: Stage-1 |
| |
| STAGE PLANS: |
| Stage: Stage-1 |
| Tez |
| #### A masked pattern was here #### |
| Vertices: |
| Map 1 |
| Map Operator Tree: |
| TableScan |
| alias: ice01 |
| branch name: branch_test1 |
| Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE |
| Select Operator |
| expressions: a (type: int), b (type: string), c (type: int) |
| outputColumnNames: _col0, _col1, _col2 |
| Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE |
| File Output Operator |
| compressed: false |
| Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE |
| table: |
| input format: org.apache.hadoop.mapred.SequenceFileInputFormat |
| output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat |
| serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| Execution mode: vectorized |
| |
| Stage: Stage-0 |
| Fetch Operator |
| limit: -1 |
| Processor Tree: |
| ListSink |
| |
| PREHOOK: query: select * from default.ice01.branch_test1 |
| PREHOOK: type: QUERY |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: hdfs://### HDFS PATH ### |
| POSTHOOK: query: select * from default.ice01.branch_test1 |
| POSTHOOK: type: QUERY |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: hdfs://### HDFS PATH ### |
| 77 one 88 |
| PREHOOK: query: drop table ice01 |
| PREHOOK: type: DROPTABLE |
| PREHOOK: Input: default@ice01 |
| PREHOOK: Output: default@ice01 |
| POSTHOOK: query: drop table ice01 |
| POSTHOOK: type: DROPTABLE |
| POSTHOOK: Input: default@ice01 |
| POSTHOOK: Output: default@ice01 |
| PREHOOK: query: drop table source01 |
| PREHOOK: type: DROPTABLE |
| PREHOOK: Input: default@source01 |
| PREHOOK: Output: default@source01 |
| POSTHOOK: query: drop table source01 |
| POSTHOOK: type: DROPTABLE |
| POSTHOOK: Input: default@source01 |
| POSTHOOK: Output: default@source01 |