blob: a839649ede54932df43a566be7651d0445a7b64b [file] [log] [blame]
PREHOOK: query: DROP TABLE table1
PREHOOK: type: DROPTABLE
POSTHOOK: query: DROP TABLE table1
POSTHOOK: type: DROPTABLE
#### A masked pattern was here ####
PREHOOK: type: CREATETABLE
PREHOOK: Input: ### test.blobstore.path ###/table1
PREHOOK: Output: database:default
PREHOOK: Output: default@table1
#### A masked pattern was here ####
POSTHOOK: type: CREATETABLE
POSTHOOK: Input: ### test.blobstore.path ###/table1
POSTHOOK: Output: database:default
POSTHOOK: Output: default@table1
PREHOOK: query: INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1
POSTHOOK: query: INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1
POSTHOOK: Output: default@table1@key=101
POSTHOOK: Output: default@table1@key=202
POSTHOOK: Output: default@table1@key=303
POSTHOOK: Output: default@table1@key=404
POSTHOOK: Output: default@table1@key=505
POSTHOOK: Lineage: table1 PARTITION(key=101).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=202).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=303).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=404).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=505).id SCRIPT []
PREHOOK: query: INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1
POSTHOOK: query: INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1
POSTHOOK: Output: default@table1@key=101
POSTHOOK: Output: default@table1@key=202
POSTHOOK: Output: default@table1@key=303
POSTHOOK: Output: default@table1@key=404
POSTHOOK: Output: default@table1@key=505
POSTHOOK: Lineage: table1 PARTITION(key=101).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=202).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=303).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=404).id SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(key=505).id SCRIPT []
PREHOOK: query: SELECT * FROM table1
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
PREHOOK: Input: default@table1@key=101
PREHOOK: Input: default@table1@key=202
PREHOOK: Input: default@table1@key=303
PREHOOK: Input: default@table1@key=404
PREHOOK: Input: default@table1@key=505
#### A masked pattern was here ####
POSTHOOK: query: SELECT * FROM table1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
POSTHOOK: Input: default@table1@key=101
POSTHOOK: Input: default@table1@key=202
POSTHOOK: Input: default@table1@key=303
POSTHOOK: Input: default@table1@key=404
POSTHOOK: Input: default@table1@key=505
#### A masked pattern was here ####
1 101
1 101
2 202
2 202
3 303
3 303
4 404
4 404
5 505
5 505
PREHOOK: query: EXPLAIN EXTENDED INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1
POSTHOOK: query: EXPLAIN EXTENDED INSERT INTO TABLE table1 PARTITION (key) VALUES (1, '101'), (2, '202'), (3, '303'), (4, '404'), (5, '505')
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1
STAGE DEPENDENCIES:
Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
Stage-2 depends on stages: Stage-0, Stage-3
Stage-3 depends on stages: Stage-1
STAGE PLANS:
Stage: Stage-1
Map Reduce
Map Operator Tree:
TableScan
alias: _dummy_table
Row Limit Per Split: 1
Statistics: Num rows: 1 Data size: 10 Basic stats: COMPLETE Column stats: COMPLETE
GatherStats: false
Select Operator
expressions: array(const struct(1,'101'),const struct(2,'202'),const struct(3,'303'),const struct(4,'404'),const struct(5,'505')) (type: array<struct<col1:int,col2:string>>)
outputColumnNames: _col0
Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: COMPLETE
UDTF Operator
Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: COMPLETE
function name: inline
Select Operator
expressions: col1 (type: int), col2 (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
bucketingVersion: 2
key expressions: _col1 (type: string), _bucket_number (type: string), _col0 (type: int)
null sort order: aaa
numBuckets: 2
sort order: +++
Map-reduce partition columns: _col1 (type: string)
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
tag: -1
auto parallelism: false
Select Operator
expressions: _col0 (type: int), _col1 (type: string)
outputColumnNames: id, key
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: min(id), max(id), count(1), count(id), compute_bit_vector(id, 'hll')
keys: key (type: string)
minReductionHashAggr: 0.99
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
Statistics: Num rows: 1 Data size: 352 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
bucketingVersion: 1
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types string,int,int,bigint,bigint,binary
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: dummy_path
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version 2
column.name.delimiter ,
columns
columns.comments
columns.types
#### A masked pattern was here ####
name _dummy_database._dummy_table
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
name: _dummy_database._dummy_table
name: _dummy_database._dummy_table
Truncated Path -> Alias:
#### A masked pattern was here ####
Needs Tagging: false
Reduce Operator Tree:
Select Operator
expressions: KEY._col0 (type: int), KEY._col1 (type: string), KEY._bucket_number (type: string)
outputColumnNames: _col0, _col1, _bucket_number
File Output Operator
bucketingVersion: 2
compressed: false
GlobalTableId: 1
directory: ### BLOBSTORE_STAGING_PATH ###
Dp Sort State: PARTITION_BUCKET_SORTED
NumFilesPerFileSink: 1
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE
Stats Publishing Key Prefix: ### BLOBSTORE_STAGING_PATH ###
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucket_count 2
bucket_field_name id
bucketing_version 2
column.name.delimiter ,
columns id
columns.comments
columns.types int
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
partition_columns key
partition_columns.types string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
TotalFiles: 1
GatherStats: true
MultiFileSpray: false
Stage: Stage-0
Move Operator
tables:
partition:
key
replace: false
source: ### BLOBSTORE_STAGING_PATH ###
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucket_count 2
bucket_field_name id
bucketing_version 2
column.name.delimiter ,
columns id
columns.comments
columns.types int
#### A masked pattern was here ####
location ### test.blobstore.path ###/table1
name default.table1
partition_columns key
partition_columns.types string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.table1
Stage: Stage-2
Stats Work
Basic Stats Work:
Stats Aggregation Key Prefix: ### BLOBSTORE_STAGING_PATH ###
Column Stats Desc:
Columns: id
Column Types: int
Table: default.table1
Is Table Level Stats: false
Stage: Stage-3
Map Reduce
Map Operator Tree:
TableScan
GatherStats: false
Reduce Output Operator
bucketingVersion: 2
key expressions: _col0 (type: string)
null sort order: z
numBuckets: -1
sort order: +
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 1 Data size: 352 Basic stats: COMPLETE Column stats: COMPLETE
tag: -1
value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: binary)
auto parallelism: false
Execution mode: vectorized
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: -mr-10002
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types string,int,int,bigint,bigint,binary
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3,_col4,_col5
columns.types string,int,int,bigint,bigint,binary
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Truncated Path -> Alias:
#### A masked pattern was here ####
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
aggregations: min(VALUE._col0), max(VALUE._col1), count(VALUE._col2), count(VALUE._col3), compute_bit_vector(VALUE._col4)
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
Statistics: Num rows: 1 Data size: 352 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: 'LONG' (type: string), UDFToLong(_col1) (type: bigint), UDFToLong(_col2) (type: bigint), (_col3 - _col4) (type: bigint), COALESCE(ndv_compute_bit_vector(_col5),0) (type: bigint), _col5 (type: binary), _col0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
Statistics: Num rows: 1 Data size: 448 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
bucketingVersion: 2
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
Statistics: Num rows: 1 Data size: 448 Basic stats: COMPLETE Column stats: COMPLETE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
bucketing_version -1
columns _col0,_col1,_col2,_col3,_col4,_col5,_col6
columns.types string:bigint:bigint:bigint:bigint:binary:string
escape.delim \
hive.serialization.extend.additional.nesting.levels true
serialization.escape.crlf true
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
PREHOOK: query: DROP TABLE table1
PREHOOK: type: DROPTABLE
PREHOOK: Input: default@table1
PREHOOK: Output: default@table1
POSTHOOK: query: DROP TABLE table1
POSTHOOK: type: DROPTABLE
POSTHOOK: Input: default@table1
POSTHOOK: Output: default@table1
PREHOOK: query: CREATE TABLE table1 (name string, age int) PARTITIONED BY (country string, state string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@table1
POSTHOOK: query: CREATE TABLE table1 (name string, age int) PARTITIONED BY (country string, state string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@table1
PREHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='CA') values ('John Doe', 23), ('Jane Doe', 22)
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1@country=USA/state=CA
POSTHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='CA') values ('John Doe', 23), ('Jane Doe', 22)
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1@country=USA/state=CA
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=CA).age SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=CA).name SCRIPT []
PREHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='CA') values ('Mark Cage', 38), ('Mirna Cage', 37)
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1@country=USA/state=CA
POSTHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='CA') values ('Mark Cage', 38), ('Mirna Cage', 37)
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1@country=USA/state=CA
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=CA).age SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=CA).name SCRIPT []
PREHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='TX') values ('Bill Rose', 52), ('Maria Full', 50)
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table1@country=USA/state=TX
POSTHOOK: query: INSERT INTO table1 PARTITION (country='USA', state='TX') values ('Bill Rose', 52), ('Maria Full', 50)
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table1@country=USA/state=TX
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=TX).age SCRIPT []
POSTHOOK: Lineage: table1 PARTITION(country=USA,state=TX).name SCRIPT []
#### A masked pattern was here ####
PREHOOK: type: CREATETABLE
PREHOOK: Input: ### test.blobstore.path ###/table2
PREHOOK: Output: database:default
PREHOOK: Output: default@table2
#### A masked pattern was here ####
POSTHOOK: type: CREATETABLE
POSTHOOK: Input: ### test.blobstore.path ###/table2
POSTHOOK: Output: database:default
POSTHOOK: Output: default@table2
PREHOOK: query: INSERT INTO TABLE table2 PARTITION (country, state) SELECT * FROM table1
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
PREHOOK: Input: default@table1@country=USA/state=CA
PREHOOK: Input: default@table1@country=USA/state=TX
PREHOOK: Output: default@table2
POSTHOOK: query: INSERT INTO TABLE table2 PARTITION (country, state) SELECT * FROM table1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
POSTHOOK: Input: default@table1@country=USA/state=CA
POSTHOOK: Input: default@table1@country=USA/state=TX
POSTHOOK: Output: default@table2
POSTHOOK: Output: default@table2@country=USA/state=CA
POSTHOOK: Output: default@table2@country=USA/state=TX
POSTHOOK: Lineage: table2 PARTITION(country=USA,state=CA).age SIMPLE [(table1)table1.FieldSchema(name:age, type:int, comment:null), ]
POSTHOOK: Lineage: table2 PARTITION(country=USA,state=CA).name SIMPLE [(table1)table1.FieldSchema(name:name, type:string, comment:null), ]
POSTHOOK: Lineage: table2 PARTITION(country=USA,state=TX).age SIMPLE [(table1)table1.FieldSchema(name:age, type:int, comment:null), ]
POSTHOOK: Lineage: table2 PARTITION(country=USA,state=TX).name SIMPLE [(table1)table1.FieldSchema(name:name, type:string, comment:null), ]
PREHOOK: query: INSERT INTO TABLE table2 PARTITION (country='MEX', state) VALUES ('Peter Mo', 87, 'SON')
PREHOOK: type: QUERY
PREHOOK: Input: _dummy_database@_dummy_table
PREHOOK: Output: default@table2@country=MEX
POSTHOOK: query: INSERT INTO TABLE table2 PARTITION (country='MEX', state) VALUES ('Peter Mo', 87, 'SON')
POSTHOOK: type: QUERY
POSTHOOK: Input: _dummy_database@_dummy_table
POSTHOOK: Output: default@table2@country=MEX/state=SON
POSTHOOK: Lineage: table2 PARTITION(country=MEX,state=SON).age SCRIPT []
POSTHOOK: Lineage: table2 PARTITION(country=MEX,state=SON).name SCRIPT []
PREHOOK: query: SHOW PARTITIONS table2
PREHOOK: type: SHOWPARTITIONS
PREHOOK: Input: default@table2
POSTHOOK: query: SHOW PARTITIONS table2
POSTHOOK: type: SHOWPARTITIONS
POSTHOOK: Input: default@table2
country=MEX/state=SON
country=USA/state=CA
country=USA/state=TX
PREHOOK: query: SELECT * FROM table2
PREHOOK: type: QUERY
PREHOOK: Input: default@table2
PREHOOK: Input: default@table2@country=MEX/state=SON
PREHOOK: Input: default@table2@country=USA/state=CA
PREHOOK: Input: default@table2@country=USA/state=TX
#### A masked pattern was here ####
POSTHOOK: query: SELECT * FROM table2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table2
POSTHOOK: Input: default@table2@country=MEX/state=SON
POSTHOOK: Input: default@table2@country=USA/state=CA
POSTHOOK: Input: default@table2@country=USA/state=TX
#### A masked pattern was here ####
Bill Rose 52 USA TX
Jane Doe 22 USA CA
John Doe 23 USA CA
Maria Full 50 USA TX
Mark Cage 38 USA CA
Mirna Cage 37 USA CA
Peter Mo 87 MEX SON
PREHOOK: query: DROP TABLE table2
PREHOOK: type: DROPTABLE
PREHOOK: Input: default@table2
PREHOOK: Output: default@table2
POSTHOOK: query: DROP TABLE table2
POSTHOOK: type: DROPTABLE
POSTHOOK: Input: default@table2
POSTHOOK: Output: default@table2
PREHOOK: query: DROP TABLE table1
PREHOOK: type: DROPTABLE
PREHOOK: Input: default@table1
PREHOOK: Output: default@table1
POSTHOOK: query: DROP TABLE table1
POSTHOOK: type: DROPTABLE
POSTHOOK: Input: default@table1
POSTHOOK: Output: default@table1