blob: 30a8d9f2d459050898c327ec15fbfd81347e2363 [file] [log] [blame]
PREHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@dest1_n84
POSTHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@dest1_n84
PREHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@dest2_n20
POSTHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@dest2_n20
PREHOOK: query: EXPLAIN EXTENDED
FROM srcpart
INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12'
INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12'
PREHOOK: type: QUERY
POSTHOOK: query: EXPLAIN EXTENDED
FROM srcpart
INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12'
INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12'
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-2 is a root stage
Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6
Stage-5
Stage-0 depends on stages: Stage-5, Stage-4, Stage-7
Stage-3 depends on stages: Stage-0, Stage-10
Stage-4
Stage-6
Stage-7 depends on stages: Stage-6
Stage-1 depends on stages: Stage-2
Stage-9 depends on stages: Stage-1, Stage-10
Stage-10 depends on stages: Stage-2
STAGE PLANS:
Stage: Stage-2
Map Reduce
Map Operator Tree:
TableScan
alias: srcpart
Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: ((ds = '2008-04-08') and (key < 100)) (type: boolean)
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 1
#### A masked pattern was here ####
NumFilesPerFileSink: 1
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
TotalFiles: 1
GatherStats: true
MultiFileSpray: false
Select Operator
expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-08' (type: string)
outputColumnNames: key, value, hr, ds
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll')
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
null sort order:
sort order:
Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
tag: -1
value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col3 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
auto parallelism: false
Filter Operator
isSamplingPred: false
predicate: ((ds = '2008-04-09') and (key < 100)) (type: boolean)
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-09' (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 2
#### A masked pattern was here ####
NumFilesPerFileSink: 1
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest2_n20
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest2_n20
TotalFiles: 1
GatherStats: true
MultiFileSpray: false
Select Operator
expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-09' (type: string)
outputColumnNames: key, value, hr, ds
Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll')
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3
columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-08
hr 12
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
numFiles 1
numRows 500
partition_columns ds/hr
partition_columns.types string:string
rawDataSize 5312
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 5812
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
#### A masked pattern was here ####
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-09
hr 12
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
numFiles 1
numRows 500
partition_columns ds/hr
partition_columns.types string:string
rawDataSize 5312
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 5812
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
/srcpart/ds=2008-04-08/hr=12 [srcpart]
/srcpart/ds=2008-04-09/hr=12 [srcpart]
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
columns _col0,_col1,_col2,_col3
columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>
escape.delim \
hive.serialization.extend.additional.nesting.levels true
serialization.escape.crlf true
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Stage: Stage-8
Conditional Operator
Stage: Stage-5
Move Operator
files:
hdfs directory: true
#### A masked pattern was here ####
Stage: Stage-0
Move Operator
tables:
replace: true
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
Stage: Stage-3
Stats Work
Basic Stats Work:
#### A masked pattern was here ####
Column Stats Desc:
Columns: key, value, hr, ds
Column Types: int, string, string, string
Table: default.dest1_n84
Is Table Level Stats: true
Stage: Stage-4
Map Reduce
Map Operator Tree:
TableScan
GatherStats: false
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: -ext-10004
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
name: default.dest1_n84
Truncated Path -> Alias:
#### A masked pattern was here ####
Stage: Stage-6
Map Reduce
Map Operator Tree:
TableScan
GatherStats: false
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: -ext-10004
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1_n84
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1_n84
name: default.dest1_n84
Truncated Path -> Alias:
#### A masked pattern was here ####
Stage: Stage-7
Move Operator
files:
hdfs directory: true
#### A masked pattern was here ####
Stage: Stage-1
Move Operator
tables:
replace: true
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}}
bucket_count -1
bucketing_version 2
column.name.delimiter ,
columns key,value,hr,ds
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest2_n20
numFiles 0
numRows 0
rawDataSize 0
serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest2_n20
Stage: Stage-9
Stats Work
Basic Stats Work:
#### A masked pattern was here ####
Column Stats Desc:
Columns: key, value, hr, ds
Column Types: int, string, string, string
Table: default.dest2_n20
Is Table Level Stats: true
Stage: Stage-10
Map Reduce
Map Operator Tree:
TableScan
GatherStats: false
Reduce Output Operator
null sort order:
sort order:
Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
tag: -1
value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col3 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
auto parallelism: false
Execution mode: vectorized
Path -> Alias:
#### A masked pattern was here ####
Path -> Partition:
#### A masked pattern was here ####
Partition
base file name: -mr-10005
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3
columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
column.name.delimiter ,
columns _col0,_col1,_col2,_col3
columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>,struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>
escape.delim \
serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Truncated Path -> Alias:
#### A masked pattern was here ####
Needs Tagging: false
Reduce Operator Tree:
Group By Operator
aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
columns _col0,_col1,_col2,_col3
columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>
escape.delim \
hive.serialization.extend.additional.nesting.levels true
serialization.escape.crlf true
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
PREHOOK: query: FROM srcpart
INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12'
INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12'
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
PREHOOK: Output: default@dest1_n84
PREHOOK: Output: default@dest2_n20
POSTHOOK: query: FROM srcpart
INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12'
INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
POSTHOOK: Output: default@dest1_n84
POSTHOOK: Output: default@dest2_n20
POSTHOOK: Lineage: dest1_n84.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
POSTHOOK: Lineage: dest1_n84.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
POSTHOOK: Lineage: dest1_n84.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1_n84.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest2_n20.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
POSTHOOK: Lineage: dest2_n20.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
POSTHOOK: Lineage: dest2_n20.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest2_n20.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
PREHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1_n84
#### A masked pattern was here ####
POSTHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1_n84
#### A masked pattern was here ####
0 val_0 12 2008-04-08
0 val_0 12 2008-04-08
0 val_0 12 2008-04-08
10 val_10 12 2008-04-08
11 val_11 12 2008-04-08
12 val_12 12 2008-04-08
12 val_12 12 2008-04-08
15 val_15 12 2008-04-08
15 val_15 12 2008-04-08
17 val_17 12 2008-04-08
18 val_18 12 2008-04-08
18 val_18 12 2008-04-08
19 val_19 12 2008-04-08
2 val_2 12 2008-04-08
20 val_20 12 2008-04-08
24 val_24 12 2008-04-08
24 val_24 12 2008-04-08
26 val_26 12 2008-04-08
26 val_26 12 2008-04-08
27 val_27 12 2008-04-08
28 val_28 12 2008-04-08
30 val_30 12 2008-04-08
33 val_33 12 2008-04-08
34 val_34 12 2008-04-08
35 val_35 12 2008-04-08
35 val_35 12 2008-04-08
35 val_35 12 2008-04-08
37 val_37 12 2008-04-08
37 val_37 12 2008-04-08
4 val_4 12 2008-04-08
41 val_41 12 2008-04-08
42 val_42 12 2008-04-08
42 val_42 12 2008-04-08
43 val_43 12 2008-04-08
44 val_44 12 2008-04-08
47 val_47 12 2008-04-08
5 val_5 12 2008-04-08
5 val_5 12 2008-04-08
5 val_5 12 2008-04-08
51 val_51 12 2008-04-08
51 val_51 12 2008-04-08
53 val_53 12 2008-04-08
54 val_54 12 2008-04-08
57 val_57 12 2008-04-08
58 val_58 12 2008-04-08
58 val_58 12 2008-04-08
64 val_64 12 2008-04-08
65 val_65 12 2008-04-08
66 val_66 12 2008-04-08
67 val_67 12 2008-04-08
67 val_67 12 2008-04-08
69 val_69 12 2008-04-08
70 val_70 12 2008-04-08
70 val_70 12 2008-04-08
70 val_70 12 2008-04-08
72 val_72 12 2008-04-08
72 val_72 12 2008-04-08
74 val_74 12 2008-04-08
76 val_76 12 2008-04-08
76 val_76 12 2008-04-08
77 val_77 12 2008-04-08
78 val_78 12 2008-04-08
8 val_8 12 2008-04-08
80 val_80 12 2008-04-08
82 val_82 12 2008-04-08
83 val_83 12 2008-04-08
83 val_83 12 2008-04-08
84 val_84 12 2008-04-08
84 val_84 12 2008-04-08
85 val_85 12 2008-04-08
86 val_86 12 2008-04-08
87 val_87 12 2008-04-08
9 val_9 12 2008-04-08
90 val_90 12 2008-04-08
90 val_90 12 2008-04-08
90 val_90 12 2008-04-08
92 val_92 12 2008-04-08
95 val_95 12 2008-04-08
95 val_95 12 2008-04-08
96 val_96 12 2008-04-08
97 val_97 12 2008-04-08
97 val_97 12 2008-04-08
98 val_98 12 2008-04-08
98 val_98 12 2008-04-08
PREHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr
PREHOOK: type: QUERY
PREHOOK: Input: default@dest2_n20
#### A masked pattern was here ####
POSTHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest2_n20
#### A masked pattern was here ####
0 val_0 12 2008-04-09
0 val_0 12 2008-04-09
0 val_0 12 2008-04-09
10 val_10 12 2008-04-09
11 val_11 12 2008-04-09
12 val_12 12 2008-04-09
12 val_12 12 2008-04-09
15 val_15 12 2008-04-09
15 val_15 12 2008-04-09
17 val_17 12 2008-04-09
18 val_18 12 2008-04-09
18 val_18 12 2008-04-09
19 val_19 12 2008-04-09
2 val_2 12 2008-04-09
20 val_20 12 2008-04-09
24 val_24 12 2008-04-09
24 val_24 12 2008-04-09
26 val_26 12 2008-04-09
26 val_26 12 2008-04-09
27 val_27 12 2008-04-09
28 val_28 12 2008-04-09
30 val_30 12 2008-04-09
33 val_33 12 2008-04-09
34 val_34 12 2008-04-09
35 val_35 12 2008-04-09
35 val_35 12 2008-04-09
35 val_35 12 2008-04-09
37 val_37 12 2008-04-09
37 val_37 12 2008-04-09
4 val_4 12 2008-04-09
41 val_41 12 2008-04-09
42 val_42 12 2008-04-09
42 val_42 12 2008-04-09
43 val_43 12 2008-04-09
44 val_44 12 2008-04-09
47 val_47 12 2008-04-09
5 val_5 12 2008-04-09
5 val_5 12 2008-04-09
5 val_5 12 2008-04-09
51 val_51 12 2008-04-09
51 val_51 12 2008-04-09
53 val_53 12 2008-04-09
54 val_54 12 2008-04-09
57 val_57 12 2008-04-09
58 val_58 12 2008-04-09
58 val_58 12 2008-04-09
64 val_64 12 2008-04-09
65 val_65 12 2008-04-09
66 val_66 12 2008-04-09
67 val_67 12 2008-04-09
67 val_67 12 2008-04-09
69 val_69 12 2008-04-09
70 val_70 12 2008-04-09
70 val_70 12 2008-04-09
70 val_70 12 2008-04-09
72 val_72 12 2008-04-09
72 val_72 12 2008-04-09
74 val_74 12 2008-04-09
76 val_76 12 2008-04-09
76 val_76 12 2008-04-09
77 val_77 12 2008-04-09
78 val_78 12 2008-04-09
8 val_8 12 2008-04-09
80 val_80 12 2008-04-09
82 val_82 12 2008-04-09
83 val_83 12 2008-04-09
83 val_83 12 2008-04-09
84 val_84 12 2008-04-09
84 val_84 12 2008-04-09
85 val_85 12 2008-04-09
86 val_86 12 2008-04-09
87 val_87 12 2008-04-09
9 val_9 12 2008-04-09
90 val_90 12 2008-04-09
90 val_90 12 2008-04-09
90 val_90 12 2008-04-09
92 val_92 12 2008-04-09
95 val_95 12 2008-04-09
95 val_95 12 2008-04-09
96 val_96 12 2008-04-09
97 val_97 12 2008-04-09
97 val_97 12 2008-04-09
98 val_98 12 2008-04-09
98 val_98 12 2008-04-09