blob: d09ee89b783c6823a011666dc809b51e75649ab9 [file] [log] [blame]
PREHOOK: query: EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
USING 'cat' AS (ds, tkey, tvalue)
CLUSTER BY tkey
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
PREHOOK: type: QUERY
PREHOOK: Input: cat
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
PREHOOK: Output: hdfs://### HDFS PATH ###
POSTHOOK: query: EXPLAIN EXTENDED
FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
USING 'cat' AS (ds, tkey, tvalue)
CLUSTER BY tkey
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
POSTHOOK: type: QUERY
POSTHOOK: Input: cat
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
POSTHOOK: Output: hdfs://### HDFS PATH ###
STAGE DEPENDENCIES:
Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
Stage: Stage-1
Tez
#### A masked pattern was here ####
Edges:
Reducer 2 <- Map 1 (SIMPLE_EDGE)
#### A masked pattern was here ####
Vertices:
Map 1
Map Operator Tree:
TableScan
alias: src
Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE
GatherStats: false
Select Operator
expressions: ds (type: string), key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE
Transform Operator
command: cat
output info:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version -1
columns _col0,_col1,_col2
columns.types string,string,string
field.delim 9
serialization.format 9
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
isSamplingPred: false
predicate: ((_col1 < 100) and (_col0 = '2008-04-08')) (type: boolean)
Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
bucketingVersion: 2
key expressions: _col1 (type: string)
null sort order: a
numBuckets: -1
sort order: +
Map-reduce partition columns: _col1 (type: string)
Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE
tag: -1
value expressions: '2008-04-08' (type: string), _col1 (type: string), _col2 (type: string)
auto parallelism: true
Path -> Alias:
hdfs://### HDFS PATH ### [src]
hdfs://### HDFS PATH ### [src]
hdfs://### HDFS PATH ### [src]
hdfs://### HDFS PATH ### [src]
Path -> Partition:
hdfs://### HDFS PATH ###
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-08
hr 11
properties:
column.name.delimiter ,
columns key,value
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
hdfs://### HDFS PATH ###
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-08
hr 12
properties:
column.name.delimiter ,
columns key,value
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
hdfs://### HDFS PATH ###
Partition
base file name: hr=11
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-09
hr 11
properties:
column.name.delimiter ,
columns key,value
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
hdfs://### HDFS PATH ###
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
partition values:
ds 2008-04-09
hr 12
properties:
column.name.delimiter ,
columns key,value
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
bucketing_version 2
column.name.delimiter ,
columns key,value
columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
location hdfs://### HDFS PATH ###
name default.srcpart
partition_columns ds/hr
partition_columns.types string:string
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.srcpart
name: default.srcpart
Truncated Path -> Alias:
/srcpart/ds=2008-04-08/hr=11 [src]
/srcpart/ds=2008-04-08/hr=12 [src]
/srcpart/ds=2008-04-09/hr=11 [src]
/srcpart/ds=2008-04-09/hr=12 [src]
Reducer 2
Execution mode: vectorized, llap
Needs Tagging: false
Reduce Operator Tree:
Select Operator
expressions: VALUE._col1 (type: string), VALUE._col2 (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
bucketingVersion: 2
compressed: false
GlobalTableId: 0
directory: hdfs://### HDFS PATH ###
NumFilesPerFileSink: 1
Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE
Stats Publishing Key Prefix: hdfs://### HDFS PATH ###
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
properties:
bucketing_version -1
columns _col0,_col1
columns.types string:string
escape.delim \
hive.serialization.extend.additional.nesting.levels true
serialization.escape.crlf true
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
TotalFiles: 1
GatherStats: false
MultiFileSpray: false
Stage: Stage-0
Fetch Operator
limit: -1
Processor Tree:
ListSink
PREHOOK: query: FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
USING 'cat' AS (ds, tkey, tvalue)
CLUSTER BY tkey
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
PREHOOK: type: QUERY
PREHOOK: Input: cat
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
PREHOOK: Output: hdfs://### HDFS PATH ###
POSTHOOK: query: FROM (
FROM srcpart src
SELECT TRANSFORM(src.ds, src.key, src.value)
USING 'cat' AS (ds, tkey, tvalue)
CLUSTER BY tkey
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
POSTHOOK: type: QUERY
POSTHOOK: Input: cat
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
POSTHOOK: Output: hdfs://### HDFS PATH ###
0 val_0
0 val_0
0 val_0
0 val_0
0 val_0
0 val_0
10 val_10
10 val_10
11 val_11
11 val_11
12 val_12
12 val_12
12 val_12
12 val_12
15 val_15
15 val_15
15 val_15
15 val_15
17 val_17
17 val_17
18 val_18
18 val_18
18 val_18
18 val_18
19 val_19
19 val_19
2 val_2
2 val_2
20 val_20
20 val_20
24 val_24
24 val_24
24 val_24
24 val_24
26 val_26
26 val_26
26 val_26
26 val_26
27 val_27
27 val_27
28 val_28
28 val_28
30 val_30
30 val_30
33 val_33
33 val_33
34 val_34
34 val_34
35 val_35
35 val_35
35 val_35
35 val_35
35 val_35
35 val_35
37 val_37
37 val_37
37 val_37
37 val_37
4 val_4
4 val_4
41 val_41
41 val_41
42 val_42
42 val_42
42 val_42
42 val_42
43 val_43
43 val_43
44 val_44
44 val_44
47 val_47
47 val_47
5 val_5
5 val_5
5 val_5
5 val_5
5 val_5
5 val_5
51 val_51
51 val_51
51 val_51
51 val_51
53 val_53
53 val_53
54 val_54
54 val_54
57 val_57
57 val_57
58 val_58
58 val_58
58 val_58
58 val_58
64 val_64
64 val_64
65 val_65
65 val_65
66 val_66
66 val_66
67 val_67
67 val_67
67 val_67
67 val_67
69 val_69
69 val_69
70 val_70
70 val_70
70 val_70
70 val_70
70 val_70
70 val_70
72 val_72
72 val_72
72 val_72
72 val_72
74 val_74
74 val_74
76 val_76
76 val_76
76 val_76
76 val_76
77 val_77
77 val_77
78 val_78
78 val_78
8 val_8
8 val_8
80 val_80
80 val_80
82 val_82
82 val_82
83 val_83
83 val_83
83 val_83
83 val_83
84 val_84
84 val_84
84 val_84
84 val_84
85 val_85
85 val_85
86 val_86
86 val_86
87 val_87
87 val_87
9 val_9
9 val_9
90 val_90
90 val_90
90 val_90
90 val_90
90 val_90
90 val_90
92 val_92
92 val_92
95 val_95
95 val_95
95 val_95
95 val_95
96 val_96
96 val_96
97 val_97
97 val_97
97 val_97
97 val_97
98 val_98
98 val_98
98 val_98
98 val_98