90eec3dd494424752d3cc025e5b3283308ccd375
[hive.git] / ql / src / test / results / clientpositive / druid_topn.q.out
1 PREHOOK: query: CREATE EXTERNAL TABLE druid_table_1
2 STORED BY 'org.apache.hadoop.hive.druid.QTestDruidStorageHandler'
3 TBLPROPERTIES ("druid.datasource" = "wikipedia")
4 PREHOOK: type: CREATETABLE
5 PREHOOK: Output: database:default
6 PREHOOK: Output: default@druid_table_1
7 POSTHOOK: query: CREATE EXTERNAL TABLE druid_table_1
8 STORED BY 'org.apache.hadoop.hive.druid.QTestDruidStorageHandler'
9 TBLPROPERTIES ("druid.datasource" = "wikipedia")
10 POSTHOOK: type: CREATETABLE
11 POSTHOOK: Output: database:default
12 POSTHOOK: Output: default@druid_table_1
13 PREHOOK: query: DESCRIBE FORMATTED druid_table_1
14 PREHOOK: type: DESCTABLE
15 PREHOOK: Input: default@druid_table_1
16 POSTHOOK: query: DESCRIBE FORMATTED druid_table_1
17 POSTHOOK: type: DESCTABLE
18 POSTHOOK: Input: default@druid_table_1
19 # col_name              data_type               comment             
20 __time                  timestamp with local time zone  from deserializer   
21 robot                   string                  from deserializer   
22 namespace               string                  from deserializer   
23 anonymous               string                  from deserializer   
24 unpatrolled             string                  from deserializer   
25 page                    string                  from deserializer   
26 language                string                  from deserializer   
27 newpage                 string                  from deserializer   
28 user                    string                  from deserializer   
29 count                   float                   from deserializer   
30 added                   float                   from deserializer   
31 delta                   float                   from deserializer   
32 variation               float                   from deserializer   
33 deleted                 float                   from deserializer   
34                  
35 # Detailed Table Information             
36 Database:               default                  
37 #### A masked pattern was here ####
38 Retention:              0                        
39 #### A masked pattern was here ####
40 Table Type:             EXTERNAL_TABLE           
41 Table Parameters:                
42         COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
43         EXTERNAL                TRUE                
44         druid.datasource        wikipedia           
45         numFiles                0                   
46         numRows                 0                   
47         rawDataSize             0                   
48         storage_handler         org.apache.hadoop.hive.druid.QTestDruidStorageHandler
49         totalSize               0                   
50 #### A masked pattern was here ####
51                  
52 # Storage Information            
53 SerDe Library:          org.apache.hadoop.hive.druid.QTestDruidSerDe     
54 InputFormat:            null                     
55 OutputFormat:           null                     
56 Compressed:             No                       
57 Num Buckets:            -1                       
58 Bucket Columns:         []                       
59 Sort Columns:           []                       
60 Storage Desc Params:             
61         serialization.format    1                   
62 PREHOOK: query: EXPLAIN
63 SELECT robot, max(added) as m, sum(variation)
64 FROM druid_table_1
65 GROUP BY robot
66 ORDER BY m DESC
67 LIMIT 100
68 PREHOOK: type: QUERY
69 POSTHOOK: query: EXPLAIN
70 SELECT robot, max(added) as m, sum(variation)
71 FROM druid_table_1
72 GROUP BY robot
73 ORDER BY m DESC
74 LIMIT 100
75 POSTHOOK: type: QUERY
76 STAGE DEPENDENCIES:
77   Stage-0 is a root stage
78
79 STAGE PLANS:
80   Stage: Stage-0
81     Fetch Operator
82       limit: -1
83       Processor Tree:
84         TableScan
85           alias: druid_table_1
86           properties:
87             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"default","dimension":"robot"}],"limitSpec":{"type":"default","limit":100,"columns":[{"dimension":"$f1","direction":"descending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
88             druid.query.type groupBy
89           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
90           Select Operator
91             expressions: robot (type: string), $f1 (type: float), $f2 (type: float)
92             outputColumnNames: _col0, _col1, _col2
93             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
94             ListSink
95
96 PREHOOK: query: EXPLAIN
97 SELECT robot, `__time`, max(added), sum(variation) as s
98 FROM druid_table_1
99 GROUP BY robot, `__time`
100 ORDER BY s DESC
101 LIMIT 100
102 PREHOOK: type: QUERY
103 POSTHOOK: query: EXPLAIN
104 SELECT robot, `__time`, max(added), sum(variation) as s
105 FROM druid_table_1
106 GROUP BY robot, `__time`
107 ORDER BY s DESC
108 LIMIT 100
109 POSTHOOK: type: QUERY
110 STAGE DEPENDENCIES:
111   Stage-0 is a root stage
112
113 STAGE PLANS:
114   Stage: Stage-0
115     Fetch Operator
116       limit: -1
117       Processor Tree:
118         TableScan
119           alias: druid_table_1
120           properties:
121             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"extraction","dimension":"__time","outputName":"extract","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","timeZone":"US/Pacific"}},{"type":"default","dimension":"robot"}],"limitSpec":{"type":"default","limit":100,"columns":[{"dimension":"$f3","direction":"descending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f2","fieldName":"added"},{"type":"doubleSum","name":"$f3","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
122             druid.query.type groupBy
123           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
124           Select Operator
125             expressions: robot (type: string), extract (type: timestamp with local time zone), $f2 (type: float), $f3 (type: float)
126             outputColumnNames: _col0, _col1, _col2, _col3
127             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
128             ListSink
129
130 PREHOOK: query: EXPLAIN
131 SELECT robot, floor_year(`__time`), max(added), sum(variation) as s
132 FROM druid_table_1
133 GROUP BY robot, floor_year(`__time`)
134 ORDER BY s DESC
135 LIMIT 10
136 PREHOOK: type: QUERY
137 POSTHOOK: query: EXPLAIN
138 SELECT robot, floor_year(`__time`), max(added), sum(variation) as s
139 FROM druid_table_1
140 GROUP BY robot, floor_year(`__time`)
141 ORDER BY s DESC
142 LIMIT 10
143 POSTHOOK: type: QUERY
144 STAGE DEPENDENCIES:
145   Stage-0 is a root stage
146
147 STAGE PLANS:
148   Stage: Stage-0
149     Fetch Operator
150       limit: -1
151       Processor Tree:
152         TableScan
153           alias: druid_table_1
154           properties:
155             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"default","dimension":"robot"},{"type":"extraction","dimension":"__time","outputName":"floor_year","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","granularity":"year","timeZone":"US/Pacific","locale":"en-US"}}],"limitSpec":{"type":"default","limit":10,"columns":[{"dimension":"$f3","direction":"descending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f2","fieldName":"added"},{"type":"doubleSum","name":"$f3","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
156             druid.query.type groupBy
157           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
158           Select Operator
159             expressions: robot (type: string), floor_year (type: timestamp with local time zone), $f2 (type: float), $f3 (type: float)
160             outputColumnNames: _col0, _col1, _col2, _col3
161             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
162             ListSink
163
164 PREHOOK: query: EXPLAIN
165 SELECT robot, floor_month(`__time`), max(added), sum(variation) as s
166 FROM druid_table_1
167 GROUP BY robot, floor_month(`__time`)
168 ORDER BY s
169 LIMIT 10
170 PREHOOK: type: QUERY
171 POSTHOOK: query: EXPLAIN
172 SELECT robot, floor_month(`__time`), max(added), sum(variation) as s
173 FROM druid_table_1
174 GROUP BY robot, floor_month(`__time`)
175 ORDER BY s
176 LIMIT 10
177 POSTHOOK: type: QUERY
178 STAGE DEPENDENCIES:
179   Stage-0 is a root stage
180
181 STAGE PLANS:
182   Stage: Stage-0
183     Fetch Operator
184       limit: -1
185       Processor Tree:
186         TableScan
187           alias: druid_table_1
188           properties:
189             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"default","dimension":"robot"},{"type":"extraction","dimension":"__time","outputName":"floor_month","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","granularity":"month","timeZone":"US/Pacific","locale":"en-US"}}],"limitSpec":{"type":"default","limit":10,"columns":[{"dimension":"$f3","direction":"ascending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f2","fieldName":"added"},{"type":"doubleSum","name":"$f3","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
190             druid.query.type groupBy
191           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
192           Select Operator
193             expressions: robot (type: string), floor_month (type: timestamp with local time zone), $f2 (type: float), $f3 (type: float)
194             outputColumnNames: _col0, _col1, _col2, _col3
195             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
196             ListSink
197
198 PREHOOK: query: EXPLAIN
199 SELECT robot, floor_month(`__time`), max(added) as m, sum(variation) as s
200 FROM druid_table_1
201 GROUP BY robot, namespace, floor_month(`__time`)
202 ORDER BY s DESC, m DESC
203 LIMIT 10
204 PREHOOK: type: QUERY
205 POSTHOOK: query: EXPLAIN
206 SELECT robot, floor_month(`__time`), max(added) as m, sum(variation) as s
207 FROM druid_table_1
208 GROUP BY robot, namespace, floor_month(`__time`)
209 ORDER BY s DESC, m DESC
210 LIMIT 10
211 POSTHOOK: type: QUERY
212 STAGE DEPENDENCIES:
213   Stage-0 is a root stage
214
215 STAGE PLANS:
216   Stage: Stage-0
217     Fetch Operator
218       limit: -1
219       Processor Tree:
220         TableScan
221           alias: druid_table_1
222           properties:
223             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"default","dimension":"robot"},{"type":"default","dimension":"namespace"},{"type":"extraction","dimension":"__time","outputName":"floor_month","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","granularity":"month","timeZone":"US/Pacific","locale":"en-US"}}],"limitSpec":{"type":"default","limit":10,"columns":[{"dimension":"$f4","direction":"descending","dimensionOrder":"numeric"},{"dimension":"$f3","direction":"descending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f3","fieldName":"added"},{"type":"doubleSum","name":"$f4","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
224             druid.query.type groupBy
225           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
226           Select Operator
227             expressions: robot (type: string), floor_month (type: timestamp with local time zone), $f3 (type: float), $f4 (type: float)
228             outputColumnNames: _col0, _col1, _col2, _col3
229             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
230             ListSink
231
232 PREHOOK: query: EXPLAIN
233 SELECT robot, floor_month(`__time`), max(added) as m, sum(variation) as s
234 FROM druid_table_1
235 GROUP BY robot, namespace, floor_month(`__time`)
236 ORDER BY robot ASC, m DESC
237 LIMIT 10
238 PREHOOK: type: QUERY
239 POSTHOOK: query: EXPLAIN
240 SELECT robot, floor_month(`__time`), max(added) as m, sum(variation) as s
241 FROM druid_table_1
242 GROUP BY robot, namespace, floor_month(`__time`)
243 ORDER BY robot ASC, m DESC
244 LIMIT 10
245 POSTHOOK: type: QUERY
246 STAGE DEPENDENCIES:
247   Stage-0 is a root stage
248
249 STAGE PLANS:
250   Stage: Stage-0
251     Fetch Operator
252       limit: -1
253       Processor Tree:
254         TableScan
255           alias: druid_table_1
256           properties:
257             druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"default","dimension":"robot"},{"type":"default","dimension":"namespace"},{"type":"extraction","dimension":"__time","outputName":"floor_month","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","granularity":"month","timeZone":"US/Pacific","locale":"en-US"}}],"limitSpec":{"type":"default","limit":10,"columns":[{"dimension":"robot","direction":"ascending","dimensionOrder":"alphanumeric"},{"dimension":"$f3","direction":"descending","dimensionOrder":"numeric"}]},"aggregations":[{"type":"doubleMax","name":"$f3","fieldName":"added"},{"type":"doubleSum","name":"$f4","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
258             druid.query.type groupBy
259           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
260           Select Operator
261             expressions: robot (type: string), floor_month (type: timestamp with local time zone), $f3 (type: float), $f4 (type: float)
262             outputColumnNames: _col0, _col1, _col2, _col3
263             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
264             ListSink
265
266 PREHOOK: query: EXPLAIN
267 SELECT robot, floor_year(`__time`), max(added), sum(variation) as s
268 FROM druid_table_1
269 WHERE robot='1'
270 GROUP BY robot, floor_year(`__time`)
271 ORDER BY s
272 LIMIT 10
273 PREHOOK: type: QUERY
274 POSTHOOK: query: EXPLAIN
275 SELECT robot, floor_year(`__time`), max(added), sum(variation) as s
276 FROM druid_table_1
277 WHERE robot='1'
278 GROUP BY robot, floor_year(`__time`)
279 ORDER BY s
280 LIMIT 10
281 POSTHOOK: type: QUERY
282 STAGE DEPENDENCIES:
283   Stage-1 is a root stage
284   Stage-0 depends on stages: Stage-1
285
286 STAGE PLANS:
287   Stage: Stage-1
288     Map Reduce
289       Map Operator Tree:
290           TableScan
291             alias: druid_table_1
292             properties:
293               druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"year","filter":{"type":"selector","dimension":"robot","value":"1"},"aggregations":[{"type":"doubleMax","name":"$f1_0","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
294               druid.query.type timeseries
295             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
296             Select Operator
297               expressions: __time (type: timestamp with local time zone), $f1_0 (type: float), $f2 (type: float)
298               outputColumnNames: _col0, _col1, _col2
299               Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
300               Reduce Output Operator
301                 key expressions: _col2 (type: float)
302                 sort order: +
303                 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
304                 TopN Hash Memory Usage: 0.1
305                 value expressions: _col0 (type: timestamp with local time zone), _col1 (type: float)
306       Reduce Operator Tree:
307         Select Operator
308           expressions: VALUE._col0 (type: timestamp with local time zone), VALUE._col1 (type: float), KEY.reducesinkkey0 (type: float)
309           outputColumnNames: _col0, _col1, _col2
310           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
311           Limit
312             Number of rows: 10
313             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
314             Select Operator
315               expressions: '1' (type: string), _col0 (type: timestamp with local time zone), _col1 (type: float), _col2 (type: float)
316               outputColumnNames: _col0, _col1, _col2, _col3
317               Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
318               File Output Operator
319                 compressed: false
320                 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
321                 table:
322                     input format: org.apache.hadoop.mapred.SequenceFileInputFormat
323                     output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
324                     serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
325
326   Stage: Stage-0
327     Fetch Operator
328       limit: -1
329       Processor Tree:
330         ListSink
331
332 PREHOOK: query: EXPLAIN
333 SELECT robot, floor_hour(`__time`), max(added) as m, sum(variation)
334 FROM druid_table_1
335 WHERE floor_hour(`__time`)
336     BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
337         AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
338 GROUP BY robot, floor_hour(`__time`)
339 ORDER BY m
340 LIMIT 100
341 PREHOOK: type: QUERY
342 POSTHOOK: query: EXPLAIN
343 SELECT robot, floor_hour(`__time`), max(added) as m, sum(variation)
344 FROM druid_table_1
345 WHERE floor_hour(`__time`)
346     BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
347         AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
348 GROUP BY robot, floor_hour(`__time`)
349 ORDER BY m
350 LIMIT 100
351 POSTHOOK: type: QUERY
352 STAGE DEPENDENCIES:
353   Stage-1 is a root stage
354   Stage-2 depends on stages: Stage-1
355   Stage-0 depends on stages: Stage-2
356
357 STAGE PLANS:
358   Stage: Stage-1
359     Map Reduce
360       Map Operator Tree:
361           TableScan
362             alias: druid_table_1
363             filterExpr: floor_hour(__time) BETWEEN 2010-01-01 00:00:00.0 US/Pacific AND 2014-01-01 00:00:00.0 US/Pacific (type: boolean)
364             properties:
365               druid.query.json {"queryType":"select","dataSource":"wikipedia","descending":false,"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"dimensions":["robot"],"metrics":["added","variation"],"granularity":"all","pagingSpec":{"threshold":16384,"fromNext":true},"context":{"druid.query.fetch":false}}
366               druid.query.type select
367             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
368             Filter Operator
369               predicate: floor_hour(__time) BETWEEN 2010-01-01 00:00:00.0 US/Pacific AND 2014-01-01 00:00:00.0 US/Pacific (type: boolean)
370               Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
371               Select Operator
372                 expressions: robot (type: string), floor_hour(__time) (type: timestamp with local time zone), added (type: float), variation (type: float)
373                 outputColumnNames: _col0, _col1, _col2, _col3
374                 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
375                 Group By Operator
376                   aggregations: max(_col2), sum(_col3)
377                   keys: _col0 (type: string), _col1 (type: timestamp with local time zone)
378                   mode: hash
379                   outputColumnNames: _col0, _col1, _col2, _col3
380                   Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
381                   Reduce Output Operator
382                     key expressions: _col0 (type: string), _col1 (type: timestamp with local time zone)
383                     sort order: ++
384                     Map-reduce partition columns: _col0 (type: string), _col1 (type: timestamp with local time zone)
385                     Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
386                     value expressions: _col2 (type: float), _col3 (type: double)
387       Reduce Operator Tree:
388         Group By Operator
389           aggregations: max(VALUE._col0), sum(VALUE._col1)
390           keys: KEY._col0 (type: string), KEY._col1 (type: timestamp with local time zone)
391           mode: mergepartial
392           outputColumnNames: _col0, _col1, _col2, _col3
393           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
394           File Output Operator
395             compressed: false
396             table:
397                 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
398                 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
399                 serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
400
401   Stage: Stage-2
402     Map Reduce
403       Map Operator Tree:
404           TableScan
405             Reduce Output Operator
406               key expressions: _col2 (type: float)
407               sort order: +
408               Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
409               TopN Hash Memory Usage: 0.1
410               value expressions: _col0 (type: string), _col1 (type: timestamp with local time zone), _col3 (type: double)
411       Reduce Operator Tree:
412         Select Operator
413           expressions: VALUE._col0 (type: string), VALUE._col1 (type: timestamp with local time zone), KEY.reducesinkkey0 (type: float), VALUE._col2 (type: double)
414           outputColumnNames: _col0, _col1, _col2, _col3
415           Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
416           Limit
417             Number of rows: 100
418             Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
419             File Output Operator
420               compressed: false
421               Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
422               table:
423                   input format: org.apache.hadoop.mapred.SequenceFileInputFormat
424                   output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
425                   serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
426
427   Stage: Stage-0
428     Fetch Operator
429       limit: 100
430       Processor Tree:
431         ListSink
432