blob: e8dabfd7cc99edee3e2958a37b43100b8778e66e [file] [log] [blame]
PREHOOK: query: drop table if exists src2_n4
PREHOOK: type: DROPTABLE
POSTHOOK: query: drop table if exists src2_n4
POSTHOOK: type: DROPTABLE
PREHOOK: query: drop table if exists src_multi1_n3
PREHOOK: type: DROPTABLE
POSTHOOK: query: drop table if exists src_multi1_n3
POSTHOOK: type: DROPTABLE
PREHOOK: query: drop table if exists src_multi2_n4
PREHOOK: type: DROPTABLE
POSTHOOK: query: drop table if exists src_multi2_n4
POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE src2_n4 as SELECT * FROM src
PREHOOK: type: CREATETABLE_AS_SELECT
PREHOOK: Input: default@src
PREHOOK: Output: database:default
PREHOOK: Output: default@src2_n4
POSTHOOK: query: CREATE TABLE src2_n4 as SELECT * FROM src
POSTHOOK: type: CREATETABLE_AS_SELECT
POSTHOOK: Input: default@src
POSTHOOK: Output: database:default
POSTHOOK: Output: default@src2_n4
POSTHOOK: Lineage: src2_n4.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: src2_n4.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
PREHOOK: query: create table src_multi1_n3 like src
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@src_multi1_n3
POSTHOOK: query: create table src_multi1_n3 like src
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@src_multi1_n3
PREHOOK: query: create table src_multi2_n4 like src
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@src_multi2_n4
POSTHOOK: query: create table src_multi2_n4 like src
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@src_multi2_n4
PREHOOK: query: explain
from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
insert overwrite table src_multi2_n4 select key, value where key > 400 order by value
PREHOOK: type: QUERY
POSTHOOK: query: explain
from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
insert overwrite table src_multi2_n4 select key, value where key > 400 order by value
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-2 is a root stage
Stage-0 depends on stages: Stage-2
Stage-3 depends on stages: Stage-0
Stage-4 depends on stages: Stage-2
Stage-1 depends on stages: Stage-4
Stage-5 depends on stages: Stage-1
STAGE PLANS:
Stage: Stage-2
Map Reduce
Map Operator Tree:
TableScan
alias: src1
Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (UDFToDouble(key) < 10.0D) (type: boolean)
Statistics: Num rows: 8 Data size: 61 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 8 Data size: 61 Basic stats: COMPLETE Column stats: NONE
Union
Statistics: Num rows: 174 Data size: 1824 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (_col0 < 150) (type: boolean)
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
Filter Operator
predicate: (_col0 > 400) (type: boolean)
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
TableScan
alias: src2_n4
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (UDFToDouble(key) > 100.0D) (type: boolean)
Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
Union
Statistics: Num rows: 174 Data size: 1824 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (_col0 < 150) (type: boolean)
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: string)
Filter Operator
predicate: (_col0 > 400) (type: boolean)
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src_multi1_n3
Select Operator
expressions: _col0 (type: string), _col1 (type: string)
outputColumnNames: key, value
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
outputColumnNames: _col0, _col1
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Move Operator
tables:
replace: true
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src_multi1_n3
Stage: Stage-3
Stats Work
Basic Stats Work:
Column Stats Desc:
Columns: key, value
Column Types: string, string
Table: default.src_multi1_n3
Stage: Stage-4
Map Reduce
Map Operator Tree:
TableScan
Reduce Output Operator
key expressions: _col1 (type: string)
sort order: +
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: string)
Execution mode: vectorized
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src_multi2_n4
Select Operator
expressions: _col0 (type: string), _col1 (type: string)
outputColumnNames: key, value
Statistics: Num rows: 58 Data size: 608 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
outputColumnNames: _col0, _col1
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-1
Move Operator
tables:
replace: true
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.src_multi2_n4
Stage: Stage-5
Stats Work
Basic Stats Work:
Column Stats Desc:
Columns: key, value
Column Types: string, string
Table: default.src_multi2_n4
PREHOOK: query: from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
insert overwrite table src_multi2_n4 select key, value where key > 400 order by value
PREHOOK: type: QUERY
PREHOOK: Input: default@src1
PREHOOK: Input: default@src2_n4
PREHOOK: Output: default@src_multi1_n3
PREHOOK: Output: default@src_multi2_n4
POSTHOOK: query: from (select * from src1 where key < 10 union all select * from src2_n4 where key > 100) s
insert overwrite table src_multi1_n3 select key, value where key < 150 order by key
insert overwrite table src_multi2_n4 select key, value where key > 400 order by value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src1
POSTHOOK: Input: default@src2_n4
POSTHOOK: Output: default@src_multi1_n3
POSTHOOK: Output: default@src_multi2_n4
POSTHOOK: Lineage: src_multi1_n3.key EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), (src2_n4)src2_n4.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: src_multi1_n3.value EXPRESSION [(src1)src1.FieldSchema(name:value, type:string, comment:default), (src2_n4)src2_n4.FieldSchema(name:value, type:string, comment:null), ]
POSTHOOK: Lineage: src_multi2_n4.key EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), (src2_n4)src2_n4.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: src_multi2_n4.value EXPRESSION [(src1)src1.FieldSchema(name:value, type:string, comment:default), (src2_n4)src2_n4.FieldSchema(name:value, type:string, comment:null), ]
PREHOOK: query: select * from src_multi1_n3
PREHOOK: type: QUERY
PREHOOK: Input: default@src_multi1_n3
#### A masked pattern was here ####
POSTHOOK: query: select * from src_multi1_n3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src_multi1_n3
#### A masked pattern was here ####
103 val_103
103 val_103
104 val_104
104 val_104
105 val_105
111 val_111
113 val_113
113 val_113
114 val_114
116 val_116
118 val_118
118 val_118
119 val_119
119 val_119
119 val_119
120 val_120
120 val_120
125 val_125
125 val_125
126 val_126
128 val_128
128 val_128
128 val_128
129 val_129
129 val_129
131 val_131
133 val_133
134 val_134
134 val_134
136 val_136
137 val_137
137 val_137
138 val_138
138 val_138
138 val_138
138 val_138
143 val_143
145 val_145
146 val_146
146 val_146
149 val_149
149 val_149
PREHOOK: query: select * from src_multi2_n4
PREHOOK: type: QUERY
PREHOOK: Input: default@src_multi2_n4
#### A masked pattern was here ####
POSTHOOK: query: select * from src_multi2_n4
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src_multi2_n4
#### A masked pattern was here ####
401 val_401
401 val_401
401 val_401
401 val_401
401 val_401
402 val_402
403 val_403
403 val_403
403 val_403
404 val_404
404 val_404
406 val_406
406 val_406
406 val_406
406 val_406
407 val_407
409 val_409
409 val_409
409 val_409
411 val_411
413 val_413
413 val_413
414 val_414
414 val_414
417 val_417
417 val_417
417 val_417
418 val_418
419 val_419
421 val_421
424 val_424
424 val_424
427 val_427
429 val_429
429 val_429
430 val_430
430 val_430
430 val_430
431 val_431
431 val_431
431 val_431
432 val_432
435 val_435
436 val_436
437 val_437
438 val_438
438 val_438
438 val_438
439 val_439
439 val_439
443 val_443
444 val_444
446 val_446
448 val_448
449 val_449
452 val_452
453 val_453
454 val_454
454 val_454
454 val_454
455 val_455
457 val_457
458 val_458
458 val_458
459 val_459
459 val_459
460 val_460
462 val_462
462 val_462
463 val_463
463 val_463
466 val_466
466 val_466
466 val_466
467 val_467
468 val_468
468 val_468
468 val_468
468 val_468
469 val_469
469 val_469
469 val_469
469 val_469
469 val_469
470 val_470
472 val_472
475 val_475
477 val_477
478 val_478
478 val_478
479 val_479
480 val_480
480 val_480
480 val_480
481 val_481
482 val_482
483 val_483
484 val_484
485 val_485
487 val_487
489 val_489
489 val_489
489 val_489
489 val_489
490 val_490
491 val_491
492 val_492
492 val_492
493 val_493
494 val_494
495 val_495
496 val_496
497 val_497
498 val_498
498 val_498
498 val_498