1 PREHOOK: query: CREATE EXTERNAL TABLE druid_table_1
2 STORED BY 'org.apache.hadoop.hive.druid.QTestDruidStorageHandler'
3 TBLPROPERTIES ("druid.datasource" = "wikipedia")
4 PREHOOK: type: CREATETABLE
5 PREHOOK: Output: database:default
6 PREHOOK: Output: default@druid_table_1
7 POSTHOOK: query: CREATE EXTERNAL TABLE druid_table_1
8 STORED BY 'org.apache.hadoop.hive.druid.QTestDruidStorageHandler'
9 TBLPROPERTIES ("druid.datasource" = "wikipedia")
10 POSTHOOK: type: CREATETABLE
11 POSTHOOK: Output: database:default
12 POSTHOOK: Output: default@druid_table_1
13 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` >= '2010-01-01 00:00:00 UTC' AND `__time` <= '2012-03-01 00:00:00 UTC' OR added <= 0
15 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` >= '2010-01-01 00:00:00 UTC' AND `__time` <= '2012-03-01 00:00:00 UTC' OR added <= 0
18 Stage-1 is a root stage
19 Stage-0 depends on stages: Stage-1
27 filterExpr: (((__time >= TIMESTAMPLOCALTZ'2009-12-31 16:00:00.0 US/Pacific') and (__time <= TIMESTAMPLOCALTZ'2012-02-29 16:00:00.0 US/Pacific')) or (added <= 0)) (type: boolean)
29 druid.query.json {"queryType":"select","dataSource":"wikipedia","descending":false,"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"dimensions":["robot","namespace","anonymous","unpatrolled","page","language","newpage","user"],"metrics":["count","added","delta","variation","deleted"],"granularity":"all","pagingSpec":{"threshold":16384,"fromNext":true},"context":{"druid.query.fetch":false}}
30 druid.query.type select
31 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
33 predicate: (((__time >= TIMESTAMPLOCALTZ'2009-12-31 16:00:00.0 US/Pacific') and (__time <= TIMESTAMPLOCALTZ'2012-02-29 16:00:00.0 US/Pacific')) or (added <= 0)) (type: boolean)
34 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
36 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
40 outputColumnNames: _col0
41 Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
42 Reduce Output Operator
44 Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
45 value expressions: _col0 (type: bigint)
48 aggregations: count(VALUE._col0)
50 outputColumnNames: _col0
51 Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
54 Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE
56 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
57 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
58 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
66 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00 UTC'
68 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00 UTC'
71 Stage-0 is a root stage
81 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"count","name":"$f0"}],"intervals":["1900-01-01T00:00:00.000Z/2010-01-01T00:00:00.001Z"],"context":{"skipEmptyBuckets":true}}
82 druid.query.type timeseries
83 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
85 expressions: $f0 (type: bigint)
86 outputColumnNames: _col0
87 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
90 PREHOOK: query: EXPLAIN
91 SELECT max(added), sum(variation)
94 POSTHOOK: query: EXPLAIN
95 SELECT max(added), sum(variation)
99 Stage-0 is a root stage
109 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"doubleMax","name":"$f0","fieldName":"added"},{"type":"doubleSum","name":"$f1","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
110 druid.query.type timeseries
111 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
113 expressions: $f0 (type: float), $f1 (type: float)
114 outputColumnNames: _col0, _col1
115 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
118 PREHOOK: query: EXPLAIN
119 SELECT `__time`, max(added), sum(variation)
123 POSTHOOK: query: EXPLAIN
124 SELECT `__time`, max(added), sum(variation)
127 POSTHOOK: type: QUERY
129 Stage-0 is a root stage
139 druid.query.json {"queryType":"groupBy","dataSource":"wikipedia","granularity":"all","dimensions":[{"type":"extraction","dimension":"__time","outputName":"extract","extractionFn":{"type":"timeFormat","format":"yyyy-MM-dd'T'HH:mm:ss.SSS'Z'","timeZone":"US/Pacific"}}],"limitSpec":{"type":"default"},"aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"]}
140 druid.query.type groupBy
141 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
143 expressions: extract (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
144 outputColumnNames: _col0, _col1, _col2
145 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
148 PREHOOK: query: EXPLAIN
149 SELECT floor_year(`__time`), max(added), sum(variation)
151 GROUP BY floor_year(`__time`)
153 POSTHOOK: query: EXPLAIN
154 SELECT floor_year(`__time`), max(added), sum(variation)
156 GROUP BY floor_year(`__time`)
157 POSTHOOK: type: QUERY
159 Stage-0 is a root stage
169 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"year","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
170 druid.query.type timeseries
171 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
173 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
174 outputColumnNames: _col0, _col1, _col2
175 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
178 PREHOOK: query: EXPLAIN
179 SELECT floor_quarter(`__time`), max(added), sum(variation)
181 GROUP BY floor_quarter(`__time`)
183 POSTHOOK: query: EXPLAIN
184 SELECT floor_quarter(`__time`), max(added), sum(variation)
186 GROUP BY floor_quarter(`__time`)
187 POSTHOOK: type: QUERY
189 Stage-0 is a root stage
199 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"quarter","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
200 druid.query.type timeseries
201 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
203 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
204 outputColumnNames: _col0, _col1, _col2
205 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
208 PREHOOK: query: EXPLAIN
209 SELECT floor_month(`__time`), max(added), sum(variation)
211 GROUP BY floor_month(`__time`)
213 POSTHOOK: query: EXPLAIN
214 SELECT floor_month(`__time`), max(added), sum(variation)
216 GROUP BY floor_month(`__time`)
217 POSTHOOK: type: QUERY
219 Stage-0 is a root stage
229 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"month","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
230 druid.query.type timeseries
231 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
233 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
234 outputColumnNames: _col0, _col1, _col2
235 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
238 PREHOOK: query: EXPLAIN
239 SELECT floor_week(`__time`), max(added), sum(variation)
241 GROUP BY floor_week(`__time`)
243 POSTHOOK: query: EXPLAIN
244 SELECT floor_week(`__time`), max(added), sum(variation)
246 GROUP BY floor_week(`__time`)
247 POSTHOOK: type: QUERY
249 Stage-0 is a root stage
259 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"week","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
260 druid.query.type timeseries
261 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
263 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
264 outputColumnNames: _col0, _col1, _col2
265 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
268 PREHOOK: query: EXPLAIN
269 SELECT floor_day(`__time`), max(added), sum(variation)
271 GROUP BY floor_day(`__time`)
273 POSTHOOK: query: EXPLAIN
274 SELECT floor_day(`__time`), max(added), sum(variation)
276 GROUP BY floor_day(`__time`)
277 POSTHOOK: type: QUERY
279 Stage-0 is a root stage
289 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"day","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
290 druid.query.type timeseries
291 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
293 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
294 outputColumnNames: _col0, _col1, _col2
295 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
298 PREHOOK: query: EXPLAIN
299 SELECT floor_hour(`__time`), max(added), sum(variation)
301 GROUP BY floor_hour(`__time`)
303 POSTHOOK: query: EXPLAIN
304 SELECT floor_hour(`__time`), max(added), sum(variation)
306 GROUP BY floor_hour(`__time`)
307 POSTHOOK: type: QUERY
309 Stage-0 is a root stage
319 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"hour","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
320 druid.query.type timeseries
321 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
323 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
324 outputColumnNames: _col0, _col1, _col2
325 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
328 PREHOOK: query: EXPLAIN
329 SELECT floor_minute(`__time`), max(added), sum(variation)
331 GROUP BY floor_minute(`__time`)
333 POSTHOOK: query: EXPLAIN
334 SELECT floor_minute(`__time`), max(added), sum(variation)
336 GROUP BY floor_minute(`__time`)
337 POSTHOOK: type: QUERY
339 Stage-0 is a root stage
349 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"minute","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
350 druid.query.type timeseries
351 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
353 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
354 outputColumnNames: _col0, _col1, _col2
355 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
358 PREHOOK: query: EXPLAIN
359 SELECT floor_second(`__time`), max(added), sum(variation)
361 GROUP BY floor_second(`__time`)
363 POSTHOOK: query: EXPLAIN
364 SELECT floor_second(`__time`), max(added), sum(variation)
366 GROUP BY floor_second(`__time`)
367 POSTHOOK: type: QUERY
369 Stage-0 is a root stage
379 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"second","aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
380 druid.query.type timeseries
381 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
383 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
384 outputColumnNames: _col0, _col1, _col2
385 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
388 PREHOOK: query: EXPLAIN
389 SELECT floor_hour(`__time`), max(added), sum(variation)
392 GROUP BY floor_hour(`__time`)
394 POSTHOOK: query: EXPLAIN
395 SELECT floor_hour(`__time`), max(added), sum(variation)
398 GROUP BY floor_hour(`__time`)
399 POSTHOOK: type: QUERY
401 Stage-0 is a root stage
411 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"hour","filter":{"type":"selector","dimension":"robot","value":"1"},"aggregations":[{"type":"doubleMax","name":"$f1","fieldName":"added"},{"type":"doubleSum","name":"$f2","fieldName":"variation"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
412 druid.query.type timeseries
413 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
415 expressions: __time (type: timestamp with local time zone), $f1 (type: float), $f2 (type: float)
416 outputColumnNames: _col0, _col1, _col2
417 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
420 PREHOOK: query: EXPLAIN
421 SELECT floor_hour(`__time`), max(added), sum(variation)
423 WHERE floor_hour(`__time`)
424 BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
425 AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
426 GROUP BY floor_hour(`__time`)
428 POSTHOOK: query: EXPLAIN
429 SELECT floor_hour(`__time`), max(added), sum(variation)
431 WHERE floor_hour(`__time`)
432 BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
433 AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
434 GROUP BY floor_hour(`__time`)
435 POSTHOOK: type: QUERY
437 Stage-1 is a root stage
438 Stage-0 depends on stages: Stage-1
446 filterExpr: floor_hour(__time) BETWEEN TIMESTAMPLOCALTZ'2010-01-01 00:00:00.0 US/Pacific' AND TIMESTAMPLOCALTZ'2014-01-01 00:00:00.0 US/Pacific' (type: boolean)
448 druid.query.json {"queryType":"select","dataSource":"wikipedia","descending":false,"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"dimensions":[],"metrics":["added","variation"],"granularity":"all","pagingSpec":{"threshold":16384,"fromNext":true},"context":{"druid.query.fetch":false}}
449 druid.query.type select
450 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
452 predicate: floor_hour(__time) BETWEEN TIMESTAMPLOCALTZ'2010-01-01 00:00:00.0 US/Pacific' AND TIMESTAMPLOCALTZ'2014-01-01 00:00:00.0 US/Pacific' (type: boolean)
453 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
455 expressions: floor_hour(__time) (type: timestamp with local time zone), added (type: float), variation (type: float)
456 outputColumnNames: _col0, _col1, _col2
457 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
459 aggregations: max(_col1), sum(_col2)
460 keys: _col0 (type: timestamp with local time zone)
462 outputColumnNames: _col0, _col1, _col2
463 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
464 Reduce Output Operator
465 key expressions: _col0 (type: timestamp with local time zone)
467 Map-reduce partition columns: _col0 (type: timestamp with local time zone)
468 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
469 value expressions: _col1 (type: float), _col2 (type: double)
470 Reduce Operator Tree:
472 aggregations: max(VALUE._col0), sum(VALUE._col1)
473 keys: KEY._col0 (type: timestamp with local time zone)
475 outputColumnNames: _col0, _col1, _col2
476 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
479 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
481 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
482 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
483 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
491 PREHOOK: query: EXPLAIN
492 SELECT subq.h, subq.m, subq.s
495 SELECT floor_hour(`__time`) as h, max(added) as m, sum(variation) as s
497 GROUP BY floor_hour(`__time`)
499 WHERE subq.h BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
500 AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
502 POSTHOOK: query: EXPLAIN
503 SELECT subq.h, subq.m, subq.s
506 SELECT floor_hour(`__time`) as h, max(added) as m, sum(variation) as s
508 GROUP BY floor_hour(`__time`)
510 WHERE subq.h BETWEEN CAST('2010-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
511 AND CAST('2014-01-01 00:00:00' AS TIMESTAMP WITH LOCAL TIME ZONE)
512 POSTHOOK: type: QUERY
514 Stage-1 is a root stage
515 Stage-0 depends on stages: Stage-1
523 filterExpr: floor_hour(__time) BETWEEN TIMESTAMPLOCALTZ'2010-01-01 00:00:00.0 US/Pacific' AND TIMESTAMPLOCALTZ'2014-01-01 00:00:00.0 US/Pacific' (type: boolean)
525 druid.query.json {"queryType":"select","dataSource":"wikipedia","descending":false,"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"dimensions":[],"metrics":["added","variation"],"granularity":"all","pagingSpec":{"threshold":16384,"fromNext":true},"context":{"druid.query.fetch":false}}
526 druid.query.type select
527 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
529 predicate: floor_hour(__time) BETWEEN TIMESTAMPLOCALTZ'2010-01-01 00:00:00.0 US/Pacific' AND TIMESTAMPLOCALTZ'2014-01-01 00:00:00.0 US/Pacific' (type: boolean)
530 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
532 expressions: floor_hour(__time) (type: timestamp with local time zone), added (type: float), variation (type: float)
533 outputColumnNames: _col0, _col1, _col2
534 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
536 aggregations: max(_col1), sum(_col2)
537 keys: _col0 (type: timestamp with local time zone)
539 outputColumnNames: _col0, _col1, _col2
540 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
541 Reduce Output Operator
542 key expressions: _col0 (type: timestamp with local time zone)
544 Map-reduce partition columns: _col0 (type: timestamp with local time zone)
545 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
546 value expressions: _col1 (type: float), _col2 (type: double)
547 Reduce Operator Tree:
549 aggregations: max(VALUE._col0), sum(VALUE._col1)
550 keys: KEY._col0 (type: timestamp with local time zone)
552 outputColumnNames: _col0, _col1, _col2
553 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
556 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
558 input format: org.apache.hadoop.mapred.SequenceFileInputFormat
559 output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
560 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
568 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1
570 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1
571 POSTHOOK: type: QUERY
573 Stage-0 is a root stage
583 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"count","name":"$f0"}],"intervals":["1900-01-01T00:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
584 druid.query.type timeseries
585 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
587 expressions: $f0 (type: bigint)
588 outputColumnNames: _col0
589 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
592 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00 UTC'
594 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00 UTC'
595 POSTHOOK: type: QUERY
597 Stage-0 is a root stage
607 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"count","name":"$f0"}],"intervals":["1900-01-01T00:00:00.000Z/2010-01-01T00:00:00.001Z"],"context":{"skipEmptyBuckets":true}}
608 druid.query.type timeseries
609 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
611 expressions: $f0 (type: bigint)
612 outputColumnNames: _col0
613 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
616 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` >= '2010-01-01 00:00:00'
618 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` >= '2010-01-01 00:00:00'
619 POSTHOOK: type: QUERY
621 Stage-0 is a root stage
631 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"count","name":"$f0"}],"intervals":["2010-01-01T08:00:00.000Z/3000-01-01T00:00:00.000Z"],"context":{"skipEmptyBuckets":true}}
632 druid.query.type timeseries
633 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
635 expressions: $f0 (type: bigint)
636 outputColumnNames: _col0
637 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
640 PREHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00' OR `__time` <= '2012-03-01 00:00:00'
642 POSTHOOK: query: EXPLAIN SELECT count(`__time`) from druid_table_1 where `__time` <= '2010-01-01 00:00:00' OR `__time` <= '2012-03-01 00:00:00'
643 POSTHOOK: type: QUERY
645 Stage-0 is a root stage
655 druid.query.json {"queryType":"timeseries","dataSource":"wikipedia","descending":false,"granularity":"all","aggregations":[{"type":"count","name":"$f0"}],"intervals":["1900-01-01T00:00:00.000Z/2012-03-01T08:00:00.001Z"],"context":{"skipEmptyBuckets":true}}
656 druid.query.type timeseries
657 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
659 expressions: $f0 (type: bigint)
660 outputColumnNames: _col0
661 Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE